Text Generation Inference
Hugging Face's high-performance text generation server
Open SourceSelf-HostedOffline CapableGPU Required (16GB+ VRAM)
0.0 (0)
About
Text Generation Inference (TGI) is a toolkit for deploying and serving Large Language Models. It includes optimized implementations for popular models and supports tensor parallelism for multi-GPU inference.
Reviews (0)
Leave a Review
No reviews yet. Be the first to review!
Details
- Category
- LLMs & Inference
- Price
- Free
- Platform
- Local/Desktop
- Difficulty
- Advanced (4/5)
- License
- Apache-2.0
- Minimum VRAM
- 16 GB
- Added
- Jan 29, 2026
Similar Tools
Featured
Run large language models locally with a simple CLI interface
Open SourceSelf-HostedOffline
Beginner
0.0 (0)
Featured
Port of Meta's LLaMA model in C/C++ for efficient CPU inference
Open SourceSelf-HostedOffline
Intermediate
0.0 (0)
Featured
High-throughput LLM serving engine with PagedAttention
Open SourceSelf-HostedOfflineGPU 16GB+
Intermediate
0.0 (0)