A high-throughput and memory-efficient inference and serving engine for LLMs
LLM inference with 7x longer context. Pure C, zero dependencies. Lossless KV cache compression + single-header library.
Faster Whisper transcription with CTranslate2
SOTA Open Source TTS
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
Translate manga/image 一键翻译各类图片内文字 https://cotrans.touhou.ai/ (no longer working)
AICI: Prompts as (Wasm) Programs
Pre-trained Neural Network models in Axon (+ 🤗 Models integration)