(1)
OpenAI’s open-weight models designed for powerful reasoning, agentic tasks
4m
100K+
42
Distilled LLaMA by DeepSeek, fast and optimized for real-world tasks
10m
100K+
76
The most advanced Qwen model yet, with major gains in text, vision, video, and reasoning.
4m
100K+
9
Efficient multimodal AI for text, image, audio, and video on low-resource devices.
9m
50K+
10
Google’s latest Gemma, in its QAT (quantization aware trained) variant
6m
50K+
21
Newest LLama 3 release with improved reasoning and generation quality
11m
50K+
18
Ministral 3: compact vision-enabled model with near-24B performance, optimized for local edge use
3m
50K+
2
Kimi K2 Thinking: open-source agent with deep reasoning, stable tool use, fast INT4, 256k context.
3m
10K+
1
DeepCoder-14B-Preview is a code reasoning LLM fine-tuned to scale up to long context lengths
11m
10K+
13
DeepSeek-V3.2 boosts efficiency and reasoning with DSA, scalable RL, agentic data—IMO/IOI wins.
3m
10K+
9
SmolLM3 is a 3.1B model for efficient on-device use, with strong performance in chat
8m
10K+
7
GLM-4.7-Flash is a top 30B-A3B MoE, balancing strong performance with efficient deployment.
2m
10K+
3
Meta’s LLama 3.1: Chat-focused, benchmark-strong, multilingual-ready.
11m
10K+
6
Efficient 80B MoE coding model with 3B activated params, 256K context, and agentic capabilities
29d
10K+
1
Qwen3 Embedding: multilingual models for advanced text/ranking tasks like retrieval & clustering.
4m
10K+
Ministral 3: compact vision-enabled model with near-24B performance, optimized for local edge use
3m
10K+
4
397B MoE model with 17B activation for reasoning, coding, agents, and multimodal understanding
22d
10K+
3
Kimi K2 Thinking: open-source agent with deep reasoning, stable tool use, fast INT4, 256k context.
3m
10K+
1
Granite Docling is a multimodal model for efficient document conversion.
5m
10K+
2
Image generation model, uses a base latent diffusion model plus a refiner.
1m
10K+
2
Safety reasoning models for policy-based text classification and foundational safety tasks.
4m
10K+
2