Displaying 1 to 30 of 63 repositories
744B MoE language model with 40B active params for reasoning, coding, and agentic tasks (FP8)
17d
4.2K
2
397B-parameter MoE multimodal LLM with 17B active params, 262K context, 201 languages
18d
3.9K
1
397B MoE model with 17B activation for reasoning, coding, agents, and multimodal understanding
18d
10K+
3
Advanced coding agent model with 80B params (3B active MoE) for code generation and debugging
25d
10K+
1
Efficient 80B MoE coding model with 3B activated params, 256K context, and agentic capabilities
25d
10K+
1
Image generation model, uses a base latent diffusion model plus a refiner.
1m
10K+
2
GLM-4.7-Flash is a top 30B-A3B MoE, balancing strong performance with efficient deployment.
1m
10K+
3
GLM-4.7-Flash is a top 30B-A3B MoE, balancing strong performance with efficient deployment.
2m
10K+
1
Devstral Small 2 is an FP8 instruct LLM for agentic SWE tasks, codebase tooling, and SWE-bench.
2m
10K+
4
FunctionGemma is a 270M open model for fine-tuned, offline function-calling agents on small devices.
2m
4.3K
1
FunctionGemma is a 270M open model for fine-tuned, offline function-calling agents on small devices.
2m
6.6K
2
Kimi K2 Thinking: open-source agent with deep reasoning, stable tool use, fast INT4, 256k context.
3m
10K+
1
Kimi K2 Thinking: open-source agent with deep reasoning, stable tool use, fast INT4, 256k context.
3m
10K+
1
DeepSeek-V3.2 boosts efficiency and reasoning with DSA, scalable RL, agentic data—IMO/IOI wins.
3m
10K+
9
Ministral 3: compact vision-enabled model with near-24B performance, optimized for local edge use
3m
10K+
4
Ministral 3: compact vision-enabled model with near-24B performance, optimized for local edge use
3m
50K+
2
Multilingual reranking model for text retrieval, scoring document relevance across 119 languages.
3m
10K+
2
Multilingual reranking model for text retrieval, scoring document relevance across 119 languages.
3m
8.9K
Snowflake’s Arctic-Embed v2.0 boosts multilingual retrieval and efficiency
4m
4.1K
Qwen3 Embedding: multilingual models for advanced text/ranking tasks like retrieval & clustering.
4m
10K+
1
Qwen3 Embedding: multilingual models for advanced text/ranking tasks like retrieval & clustering.
4m
10K+
OpenAI’s open-weight models designed for powerful reasoning, agentic tasks
4m
100K+
42
The most advanced Qwen model yet, with major gains in text, vision, video, and reasoning.
4m
100K+
9
Safety reasoning models for policy-based text classification and foundational safety tasks.
4m
10K+
2
Granite-4.0-nano: lightweight instruct model trained via SFT, RL, and merging on diverse data.
4m
8.8K
Granite-4.0-h-nano: lightweight instruct model trained via SFT, RL, and merging on diverse data.
4m
4.1K
1
OpenAI’s open-weight models designed for powerful reasoning, agentic tasks
4m
10K+
1