Overview
Create embeddings and call hosted models through the Inference API; AI Hub tries the pipeline endpoint first, then model URL fallback.
Capabilities
-
Embeddings via feature-extraction for RAG/search
-
Model override and base URL override options
-
Batch encode texts via /api/handler/execute task=encode
Setup Steps (3)
-
1
Step 1
Save token: POST /AdminApi.php {"provider":"huggingface","creds":{"api_key":"hf_…"}}.
-
2
Step 2
Encode vectors: POST /api/handler/execute {"task":"encode","texts":["hello","world"],"model":"sentence-transformers/all-mpnet-base-v2"}
-
3
Limitations
-
Embedding dimensionality depends on chosen model; verify before indexing.
-
Provider rate limits apply; use the Jobs queue for large batches.
FAQs
Can I point to a private model?
Yes, if your token has access. You can also override base URLs when needed.
Is there caching?
Yes. Deterministic requests benefit from memo-cache if configured.
Pricing
Free
Free
Great for trying the integration.
Pro
USD 9.99 / monthly
Enterprise
USD 49.99 / monthly