The open-source AI voice studio. Clone, dictate, create.
A high-throughput and memory-efficient inference and serving engine for LLMs
Go with your own intelligence - Go applications that directly integrate llama.cpp for local inference using hardware acceleration.