Use your own language models with Anam’s digital personas
API Compatibility
Streaming Support
stream: true
support - Use Server-Sent Events (SSE) for streaming chunks -
Include proper content types and formattingValidation Testing
curl
commands to compare your custom
LLM’s raw HTTP responses with those from the actual providers (OpenAI, Azure
OpenAI, or Gemini). Client libraries like the OpenAI SDK often transform
responses and extract specific values, which can mask differences in the
actual HTTP response format. Your custom implementation must match the raw
HTTP response structure, not the transformed output from client libraries.brainType
parameter is deprecated and has been replaced with llmId
to
better reflect support for custom models. For backwards compatibility, you can
pass your existing brainType
value as the llmId
and it will continue to
work.Update your code
brainType
with llmId
in your session token requests:No functional changes needed
ANAM_GPT_4O_MINI_V1
→ Works as llmId
- ANAM_LLAMA_v3_3_70B_V1
→ Works as llmId
-
CUSTOMER_CLIENT_V1
→ Works as llmId
Consider custom LLMs
LLM ID | Description | Best For |
---|---|---|
ANAM_GPT_4O_MINI_V1 | OpenAI GPT-4 Mini model | Available for backwards compatibility |
0934d97d-0c3a-4f33-91b0-5e136a0ef466 | OpenAI GPT-4.1 Mini model | Recommended for new projects |
ANAM_LLAMA_v3_3_70B_V1 | Llama 3.3 70B model | Open-source preference, larger context |
9d8900ee-257d-4401-8817-ba9c835e9d36 | Gemini 2.5 Flash model | Our fastest model |
CUSTOMER_CLIENT_V1 | Client-side LLM | When you only use .talk() commands to speak |