Tools/LLMs & Inference/Text Generation Inference

Text Generation Inference

Hugging Face's high-performance text generation server

Open SourceSelf-HostedOffline CapableGPU Required (16GB+ VRAM)
0.0 (0)

About

Text Generation Inference (TGI) is a toolkit for deploying and serving Large Language Models. It includes optimized implementations for popular models and supports tensor parallelism for multi-GPU inference.

Reviews (0)

Leave a Review

No reviews yet. Be the first to review!

Details

Price
Free
Platform
Local/Desktop
Difficulty
Advanced (4/5)
License
Apache-2.0
Minimum VRAM
16 GB
Added
Jan 29, 2026

Similar Tools

Featured

Run large language models locally with a simple CLI interface

Open SourceSelf-HostedOffline
Beginner
0.0 (0)
Featured

Port of Meta's LLaMA model in C/C++ for efficient CPU inference

Open SourceSelf-HostedOffline
Intermediate
0.0 (0)
Featured

High-throughput LLM serving engine with PagedAttention

Open SourceSelf-HostedOfflineGPU 16GB+
Intermediate
0.0 (0)