Author: darinverheijke Version: 0.0.1 Type: model
Featherless AI is a serverless inference provider that gives access to thousands of open source models such as DeepSeek, Qwen, Llama, Mistral and more.
To configure your Featherless AI provider, follow these steps:
- The Model Type field should already be set to "LLM"
- This is typically pre-selected and doesn't need to be changed
- In the Model Name field, enter the full model name exactly as it appears in Featherless AI
- You can find the list of models on https://featherless.ai/models
- Examples of available models:
deepseek-ai/DeepSeek-V3-0324mistralai/Mistral-Nemo-Instruct-2407meta-llama/Meta-Llama-3.1-8B-InstructQwen/Qwen3-32BTHUD/GLM-4-32B-0414Qwen/QwQ-32B
- In the API Key field, enter your Featherless AI API key
- If you don't have one, visit featherless.ai to get your API key
- The Completion mode should be set to "Chat"
- This is the standard mode for most language models
- Set the Model context size to the appropriate value for your chosen model
- Common values include
16384,32768 - Check the model page for the correct context size
- Set the Upper bound for max tokens
- This should typically match or be less than the context size
- Example:
4096for models with 4K context
- Click the Save button to add the model to your configuration
- Click Cancel if you want to abort the process
- Make sure you have a subscription on your Featherless AI account
- The model name must match exactly what's available in the Featherless AI catalog
- Some models may have different context sizes - check the official documentation for accurate values
- All models support chat completion mode for conversational AI applications