AI Models

Explore our comprehensive collection of AI models, each with unique capabilities and strengths.

Categories

General Purpose
Multimodal
Specialized
Code Generation
Information Retrieval

Providers

OpenAI
Anthropic
Amazon
Google
Meta
Mistral
DeepSeek
Perplexity
Qwen

Tags

new
featured
best for coding
best for chat
high context
cost effective
enterprise
research

35 Models Found

OpenAI

GPT-4o

OpenAI

OpenAI's most advanced multimodal model, capable of processing and generating text, images, and potentially other data types in real-time. It features a 128,000-token context window, delivering improved reasoning, reduced latency, and enhanced instruction-following compared to previous models. GPT-4o achieves state-of-the-art performance across benchmarks like MMLU and excels in applications requiring real-time interaction, such as conversational agents, creative writing, and multimodal analysis.

128,000 tokens
Advanced reasoning
Real-time multimodal processing
Long context handling
+2
new
featured
+1
General Purpose
View details
OpenAI

GPT-4o Mini

OpenAI

A compact, cost-efficient variant of GPT-4o, retaining 70% of its multimodal performance with a 128,000-token context window. It supports text generation, image understanding, and code generation at a fraction of the cost, making it ideal for budget-conscious applications like lightweight chatbots, content generation, and educational tools. GPT-4o Mini balances performance and affordability while maintaining strong reasoning capabilities.

128,000 tokens
Cost effectiveness
Balanced multimodal capabilities
Image understanding
+1
new
cost effective
General Purpose
View details
OpenAI

O3 Mini

OpenAI

A highly efficient, affordable model designed for everyday tasks, featuring a 32,000-token context window. O3 Mini excels in rapid text and code generation with basic reasoning capabilities, making it perfect for high-volume applications such as customer support chatbots, automated responses, and simple scripting tasks. Its low cost and fast processing speed ensure scalability for routine operations.

32,000 tokens
Speed and efficiency
Cost effectiveness
Good reasoning
new
cost effective
General Purpose
View details
Claude

Claude 3.7 Sonnet

Anthropic

Anthropic’s flagship model introducing 'visible thinking,' which displays its reasoning process in real-time for enhanced transparency. With a 200,000-token context window, it offers top-tier reasoning, coding, and multimodal capabilities (text and images). Claude 3.7 Sonnet excels in complex tasks like software development, data analysis, and problem-solving, with low hallucination rates and robust safety features, making it a trusted choice for critical applications.

200,000 tokens
Advanced reasoning
Visible thinking transparency
Superior code generation
+2
General Purpose
View details
Claude

Claude 3.5 Sonnet

Anthropic

A highly capable Anthropic model with a 200,000-token context window, offering exceptional reasoning, coding, and multimodal (text and image) performance. It includes advanced safety guardrails and improved multilingual support, making it suitable for secure, complex tasks like legal analysis, technical writing, and visual data interpretation. Claude 3.5 Sonnet is known for its low hallucination rate and precise outputs.

200,000 tokens
Strong reasoning
Low hallucination rate
Image understanding
+2
General Purpose
View details
Claude

Claude 3.5 Haiku

Anthropic

A fast, cost-effective Anthropic model with a 200,000-token context window, optimized for high-throughput, real-time applications. It delivers strong reasoning and multimodal capabilities (text and images) at a lower cost, making it ideal for tasks like customer service automation, content moderation, and quick visual analysis. Claude 3.5 Haiku balances speed, efficiency, and performance.

200,000 tokens
Speed and efficiency
Cost effectiveness
Image understanding
+1
General Purpose
View details
Claude

Claude 3.5 Haiku New

Anthropic

An upgraded version of Claude 3.5 Haiku with a 200,000-token context window, featuring improved reasoning, coding, and visual understanding. It retains the original’s speed and cost-efficiency while enhancing performance on complex prompts, making it suitable for advanced real-time applications like technical support, code review, and detailed image-based tasks.

200,000 tokens
Speed and efficiency
Cost effectiveness
Image understanding
+1
General Purpose
View details
Nova

Amazon Nova Pro

Amazon

A premium multimodal model from Amazon’s Nova family, featuring a 300,000-token context window and top-tier accuracy, speed, and cost-efficiency. It excels in processing text, images, and videos with advanced reasoning, making it ideal for agentic workflows, enterprise analytics, and complex multimodal tasks like video summarization and document processing.

300,000 tokens
Instruction following
Image and video understanding
Long context handling
+1
General Purpose
View details
Nova

Amazon Titan Premier

Amazon

A high-performance model from Amazon’s Titan family with a 128,000-token context window, designed for enterprise applications. It offers complex reasoning, enhanced guardrails, and customization options, excelling in retrieval-augmented generation (RAG) and agent-based workflows in secure environments like financial analysis and compliance monitoring.

128,000 tokens
Advanced reasoning
Enterprise applications
Long context handling
+1
General Purpose
View details
Nova

Nova Lite

Amazon

A cost-effective multimodal model from Amazon’s Nova family with a 128,000-token context window, optimized for high-throughput tasks. It supports text, image, and video processing, making it suitable for applications like customer interactions, document analysis, and lightweight visual content generation, all at a competitive price point.

128,000 tokens
Cost effectiveness
Speed and efficiency
Basic image understanding
General Purpose
View details
Nova

Nova Micro

Amazon

An ultra-efficient text-only model from Amazon’s Nova family with a 128,000-token context window, designed for low-latency, high-throughput text processing. It’s ideal for simple tasks like chatbots, text summarization, and automated responses in cost-sensitive scenarios, offering minimal overhead and maximum scalability.

128,000 tokens
Speed and efficiency
Cost effectiveness
High throughput
General Purpose
View details
Nova

Titan Express

Amazon

A fast, economical text model from Amazon’s Titan family with an 8,000-token context window, tailored for high-volume text processing. It provides reliable performance for enterprise use cases like conversational chat, text generation, and basic automation, balancing speed and cost-efficiency effectively.

8,000 tokens
Speed and efficiency
Cost effectiveness
Enterprise applications
General Purpose
View details
Nova

Titan Lite

Amazon

An ultra-lightweight text model from Amazon’s Titan family with a 4,000-token context window, offering basic text processing at the lowest cost. It’s optimized for simple, high-volume applications like short-form content generation and lightweight chatbots, prioritizing affordability and efficiency.

4,000 tokens
Cost effectiveness
Speed and efficiency
Basic reasoning
General Purpose
View details
Gemini

Gemini 2.0 Flash

Google

A fast, efficient multimodal model from Google’s Gemini 2.0 family with a 128,000-token context window, supporting text, image, and basic video understanding. It delivers strong performance for everyday tasks like content creation, image analysis, and quick coding, offering a cost-effective solution with excellent speed and scalability.

128,000 tokens
Cost effectiveness
Speed and efficiency
Image understanding
+1
General Purpose
View details
Gemini

Gemini 2.0 Pro

Google

Google’s most advanced multimodal model with an unprecedented 2,000,000-token context window, designed for complex reasoning and sophisticated multimodal tasks. It excels in deep content generation, advanced coding, and detailed image/video analysis, making it ideal for research, creative projects, and enterprise applications requiring extensive context retention.

2,000,000 tokens
Advanced reasoning
Unprecedented context window
Multimodal understanding
+1
General Purpose
View details
Gemini

Gemini 2.0 Flash Lite

Google

An ultra-efficient version of Gemini 2.0 Flash with a 32,000-token context window, optimized for maximum performance at minimal cost. It retains core multimodal capabilities (text and images), making it suitable for high-volume, cost-sensitive applications like lightweight chatbots, basic image processing, and rapid text generation.

32,000 tokens
Cost effectiveness
Speed and efficiency
Basic multimodal capabilities
General Purpose
View details
Gemini

Gemma2 9B IT

Google

An open-weight Google model with an 8,000-token context window, designed for efficiency and local deployment. It offers improved reasoning, coding, and multilingual capabilities, making it ideal for edge devices, on-device applications, and research projects needing a compact yet powerful model.

8,000 tokens
Open weights
Efficiency
Local deployment
+1
General Purpose
View details
Meta

Llama 3.3 70B

Meta

Meta’s latest open-weight model with a 128,000-token context window, delivering near-top-tier performance with lower computational demands. It features enhanced reasoning, coding, and multilingual capabilities, making it a versatile choice for high-performance applications like academic research, software development, and global content generation.

128,000 tokens
Open weights
Strong reasoning
Code generation
+1
General Purpose
View details
Meta

Llama 3.1 8B

Meta

An efficient open-weight model from Meta’s Llama 3.1 family with an 8,000-token context window, optimized for accessibility and local deployment. It provides strong performance with improved multilingual support and instruction following, ideal for cost-effective solutions like lightweight chatbots, educational tools, and on-device AI applications.

8,000 tokens
Open weights
Efficiency
Local deployment
+1
General Purpose
View details
Mistral

Mistral 8B

Mistral

An efficient Mistral AI model with a 32,000-token context window, offering strong reasoning, instruction following, and coding capabilities. Its compact size makes it suitable for a wide range of applications, including content generation, task automation, and lightweight software development, balancing performance and resource efficiency.

32,000 tokens
Efficiency
Balanced capabilities
Instruction following
General Purpose
View details
Mistral

Mistral 3B Latest

Mistral

An ultra-compact Mistral AI model with an 8,000-token context window, designed for resource-constrained environments like mobile devices and embedded systems. It delivers impressive performance for its size, excelling in lightweight tasks such as text generation, basic coding, and on-device automation with minimal latency.

8,000 tokens
Efficiency
Speed
Local deployment
+1
General Purpose
View details
Mistral

Mistral Large

Mistral

Mistral AI’s flagship model with a 32,000-token context window, offering exceptional reasoning, coding, and multilingual capabilities. Scoring 81.2% on the MMLU benchmark, it excels in complex problem-solving, creative writing, and technical tasks, making it a powerful choice for advanced applications in research, development, and global communication.

32,000 tokens
Strong reasoning
Code generation
Instruction following
+1
General Purpose
View details
Mistral

Mistral Moderation Latest

Mistral

A specialized Mistral AI model with a 32,000-token context window, purpose-built for content moderation. It detects and filters harmful or policy-violating content with high accuracy, leveraging specialized training for safety and compliance applications like social media monitoring and user-generated content review.

32,000 tokens
Content moderation
Safety filtering
Policy enforcement
Specialized
View details
Mistral

Mistral Saba Latest

Mistral

An enterprise-focused Mistral AI model with a 32,000-token context window, optimized for business applications and professional workflows. It features enhanced reasoning and document processing capabilities, making it ideal for tasks like contract analysis, report generation, and specialized knowledge extraction in corporate environments.

32,000 tokens
Enterprise applications
Strong reasoning
Instruction following
General Purpose
View details
Mistral

Mistral Small Latest

Mistral

A mid-sized Mistral AI model with a 32,000-token context window, offering a balanced mix of performance and efficiency. It provides strong reasoning, coding, and instruction-following capabilities at a reasonable cost, making it suitable for production applications like customer support automation, content creation, and moderate-scale development tasks.

32,000 tokens
Cost effectiveness
Balanced capabilities
Instruction following
General Purpose
View details
Mistral

Pixtral Large Latest

Mistral

An advanced multimodal Mistral AI model with a 32,000-token context window, specializing in image understanding and visual reasoning. It excels in complex visual tasks like detailed image analysis, object recognition, and multimodal problem-solving, making it ideal for applications in design, education, and visual data interpretation.

32,000 tokens
Image understanding
Visual reasoning
Multimodal capabilities
General Purpose
View details
Mistral

Codestral

Mistral

A dedicated code generation model from Mistral AI with a 32,000-token context window, designed for software development and debugging. It excels in code completion, testing, and generation across multiple programming languages, offering enhanced contextual understanding for tasks like automated coding, script optimization, and developer assistance.

32,000 tokens
Superior code generation
Multi-language support
Instruction following
Code Generation
View details
DeepSeek

Deepseek R1 Llama 70B

DeepSeek

A reasoning-focused DeepSeek model with a 128,000-token context window, built on the Llama architecture. It offers advanced logical reasoning and problem-solving capabilities, excelling in complex task automation, mathematical analysis, and reasoning benchmarks, making it a strong choice for research and technical applications.

128,000 tokens
Advanced reasoning
Code generation
Instruction following
General Purpose
View details
DeepSeek

Deepseek R1 Qwen 32B

DeepSeek

A mid-sized DeepSeek reasoning model with a 128,000-token context window, based on the Qwen architecture. It provides strong logical reasoning and complex task completion with lower computational demands, making it ideal for efficient analytical applications like data interpretation, automated reasoning, and technical problem-solving.

128,000 tokens
Reasoning capabilities
Efficiency
Instruction following
General Purpose
View details
Mistral

Open Mistral Nemo

Mistral

An open-source Mistral model with a 32,000-token context window, optimized with NVIDIA NeMo for superior performance and deployment flexibility. It features enhanced efficiency and hardware acceleration, making it suitable for high-performance inference on NVIDIA platforms, ideal for developers and researchers needing customizable, efficient AI solutions.

32,000 tokens
Open weights
Efficiency
Local deployment
General Purpose
View details
Perplexity

Perplexity Llama 3.1 Sonar

Perplexity

A specialized Perplexity model with a 128,000-token context window, built on Llama 3.1 for real-time information retrieval. It integrates web search capabilities and knowledge-intensive training, providing factually accurate responses for applications like research assistance, question answering, and current event analysis.

128,000 tokens
Information retrieval
Web search capabilities
Factual responses
General Purpose
View details
Perplexity

Perplexity R1 1776

Perplexity

An advanced Perplexity model with a 128,000-token context window, combining strong reasoning with real-time information retrieval. It excels in complex analytical tasks requiring up-to-date data, such as market research, policy analysis, and detailed problem-solving, offering a blend of logical depth and factual accuracy.

128,000 tokens
Advanced reasoning
Information retrieval
Factual responses
General Purpose
View details
Perplexity

Perplexity Sonar

Perplexity

Perplexity’s flagship model with a 128,000-token context window, featuring industry-leading online search and information synthesis. It provides real-time web access, source attribution, and high factual accuracy, making it ideal for knowledge-intensive applications like academic research, journalism, and real-time decision-making support.

128,000 tokens
Online search
Information retrieval
Factual responses
+1
General Purpose
View details
Qwen

Qwen 2.5 32B

Qwen

An advanced multilingual model from Alibaba Cloud with a 128,000-token context window, offering excellent performance across languages and tasks. It features sophisticated reasoning and code generation, making it ideal for global applications like cross-lingual content creation, technical development, and enterprise-grade automation.

128,000 tokens
Multilingual capabilities
Strong reasoning
Long context handling
+1
General Purpose
View details
Qwen

Qwen Plus

Qwen

An optimized Qwen model with a 32,000-token context window, enhanced for production-grade applications. It offers strong multilingual support and efficient processing, tailored for enterprise use cases like customer service automation, multilingual document analysis, and high-throughput workflows requiring reliable performance.

32,000 tokens
Balanced capabilities
Multilingual support
Instruction following
General Purpose
View details