Distributed
AI workloads can be distributed across many processors.
Distribution
-
Exolabs
https://exolabs.net/ -
Petals
GitHub - bigscience-workshop/petals: 🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
https://medium.com/@visrow/decentralized-distributed-llm-using-ai-in-cost-effective-and-environment-friendly-way-8c0a73ee9e6f -
AI Horde
https://stablehorde.net/
Parallel
Multiple GPUs can be used with llama.cpp now: