Get Started
Model Selection
Onyx offers the largest selection of models with shortest time to availablity on the market. All of these can be found natively within the LLM Node Model selector.
Onyx Power Rankings:
OpenAI gpt-4o
OpenAI’s flagship model has been a consisstent FM leveraged for use cases spanning Consumer Lifestyle, Healthtech, and even GovTech applications on Onyx
Claude 3.5 Sonnet
With the release of 3.5, Claude has been able to reclaim the crown of having the best performing FM.
Llama-3 80B
Llama 3 is the king of OSS models in terms of usage. Onyx customers can leverage the power of this model via API or through docker enabled deployment to the Cloud.
Open AI GPT
Model | Description | Context Window |
---|---|---|
gpt-4o | OpenAI’s most advanced, multimodal flagship model that’s cheaper and faster than GPT-4 Turbo | 128,000 Tokens |
gpt-4o-mini | OpenAI’s first SLM that’s cheaper and faster than 3.5-turbo | 128,000 Tokens |
gpt4-turbo | The latest GPT-4 Turbo model with vision capabilities. | 128,000 Tokens |
gpt-3.5-turbo | The latest GPT-3.5 Turbo model with higher accuracy at responding in requested formats and function calling | 16,385 Tokens |
Anthropic Claude
Model | Description | Context Window |
---|---|---|
claude-3.5-sonnet | OpenAI’s most advanced, multimodal flagship model that’s cheaper and faster than GPT-4 Turbo | 128,000 Tokens |
claude-3-opus | The latest GPT-4 Turbo model with vision capabilities. | 128,000 |
claude-3-sonnet | The latest GPT-3.5 Turbo model with higher accuracy at responding in requested formats and function calling | 16,385 |
claude-3-haiku | The latest GPT-3.5 Turbo model with higher accuracy at responding in requested formats and function calling | 16,385 |
Meta Llama
Model | Description | Context Window |
---|---|---|
Meta-Llama-3-70B | Meta’s most advanced OSS model. Deploy via API or host your own via docker deployment to your CSP of choice | 8,000 Tokens |
Meta-Llama-3-8B | Capable and fast model that is often fine tuned to optimize inference for hyperspecialized use cases | 8,000 tokens |
Google Gemini
Model | Description | Context Window |
---|---|---|
gemini-1.5-flash | Google’s fastest and most optimized multimodal LLM | 1,000,000 Tokens |
gemini-1.5-pro | Google’s most capable and powerful multimodal LLM boasting the largest context window to date | 2,000,000 Tokens |
If you require any model not listed above, reach out system@abstractionlabs.ai and our team will work on providing access to the model within 30 minutes of first response.