Hugging Face Model Inference
Unified workflow for running AI/ML inference across Hugging Face APIs, combining the Inference API, Inference Providers, and Text Generation Inference for NLP, vision, audio, and multimodal tasks. Used by ML engineers and AI application developers.
What You Can Do
MCP Tools
run-inference
Run inference on any Hugging Face model by model ID.
generate-text
Generate text using a language model via the Inference API.
classify-text
Classify text into predefined categories.
answer-question
Answer questions based on provided context.
summarize-text
Summarize text content.
translate-text
Translate text between languages.
fill-mask
Fill in masked tokens in text.
extract-features
Extract feature vectors from text for embeddings.
classify-image
Classify images into categories.
detect-objects
Detect objects in images.
transcribe-speech
Transcribe audio to text using automatic speech recognition.
generate-image
Generate images from text prompts.
zero-shot-classify
Classify text without predefined training labels.
compute-similarity
Compute similarity between sentences.
providers-chat-completion
Create chat completion via OpenAI-compatible multi-provider API.
providers-text-completion
Create text completion via multi-provider API.
providers-create-embeddings
Create text embeddings via multi-provider API.
providers-generate-image
Generate images via multi-provider API.
providers-transcribe
Transcribe audio via multi-provider API.
providers-text-to-speech
Convert text to speech via multi-provider API.
tgi-generate
Generate text using the TGI native endpoint.
tgi-chat-completions
Create chat completions using TGI OpenAI-compatible Messages API.
tgi-tokenize
Tokenize input text and return token IDs.
tgi-server-info
Get TGI server information and deployed model details.
list-provider-models
List models available across all inference providers.