LLM inference server with continuous batching & SSD caching for Apple Silicon — managed from the macOS menu bar - View it on GitHub
Star
0
Rank
13847748