Velaxe
AI Hub — Unified LLM Gateway, Chat, Embeddings & Jobs | Velaxe

AI Hub

Hugging Face — Integration

Create embeddings and call hosted models through the Inference API; AI Hub tries the pipeline endpoint first, then model URL fallback.

Overview

Create embeddings and call hosted models through the Inference API; AI Hub tries the pipeline endpoint first, then model URL fallback.

Capabilities

  • Embeddings via feature-extraction for RAG/search

  • Model override and base URL override options

  • Batch encode texts via /api/handler/execute task=encode

Setup Steps (3)

  1. 1

    Step 1

    Save token: POST /AdminApi.php {"provider":"huggingface","creds":{"api_key":"hf_…"}}.

  2. 2

    Step 2

    Encode vectors: POST /api/handler/execute {"task":"encode","texts":["hello","world"],"model":"sentence-transformers/all-mpnet-base-v2"}

  3. 3

    Step 3

    Persist vectors in your app’s store and attach ids for retrieval.

Limitations

  • Embedding dimensionality depends on chosen model; verify before indexing.

  • Provider rate limits apply; use the Jobs queue for large batches.

FAQs

Can I point to a private model?

Yes, if your token has access. You can also override base URLs when needed.

Is there caching?

Yes. Deterministic requests benefit from memo-cache if configured.

Pricing

Free

Free

Great for trying the integration.

Pro

USD 9.99 / monthly

Enterprise

USD 49.99 / monthly