Onyx offers the largest selection of models with shortest time to availablity on the market. All of these can be found natively within the LLM Node Model selector.
Model | Description | Context Window |
---|---|---|
gpt-4o | OpenAI’s most advanced, multimodal flagship model that’s cheaper and faster than GPT-4 Turbo | 128,000 Tokens |
gpt-4o-mini | OpenAI’s first SLM that’s cheaper and faster than 3.5-turbo | 128,000 Tokens |
gpt4-turbo | The latest GPT-4 Turbo model with vision capabilities. | 128,000 Tokens |
gpt-3.5-turbo | The latest GPT-3.5 Turbo model with higher accuracy at responding in requested formats and function calling | 16,385 Tokens |
Model | Description | Context Window |
---|---|---|
claude-3.5-sonnet | OpenAI’s most advanced, multimodal flagship model that’s cheaper and faster than GPT-4 Turbo | 128,000 Tokens |
claude-3-opus | The latest GPT-4 Turbo model with vision capabilities. | 128,000 |
claude-3-sonnet | The latest GPT-3.5 Turbo model with higher accuracy at responding in requested formats and function calling | 16,385 |
claude-3-haiku | The latest GPT-3.5 Turbo model with higher accuracy at responding in requested formats and function calling | 16,385 |
Model | Description | Context Window |
---|---|---|
Meta-Llama-3-70B | Meta’s most advanced OSS model. Deploy via API or host your own via docker deployment to your CSP of choice | 8,000 Tokens |
Meta-Llama-3-8B | Capable and fast model that is often fine tuned to optimize inference for hyperspecialized use cases | 8,000 tokens |
Model | Description | Context Window |
---|---|---|
gemini-1.5-flash | Google’s fastest and most optimized multimodal LLM | 1,000,000 Tokens |
gemini-1.5-pro | Google’s most capable and powerful multimodal LLM boasting the largest context window to date | 2,000,000 Tokens |