tgi

TGI

Text Generation Inference for LLMs.

TGI server and sharding configs for serving LLMs with tensor parallelism.

Configuration Recipes0

No recipes yet for TGI. Check back soon.