From ffb977241c5b68b04439e8674b6b968f4a11f55f Mon Sep 17 00:00:00 2001 From: jlarson4 Date: Thu, 9 Apr 2026 17:09:28 -0500 Subject: [PATCH] Creating a relevancy testing system for architecture gaps --- tests/unit/tools/test_relevancy.py | 234 + .../data/architecture_gaps.json | 5035 +++-- .../model_registry/data/supported_models.json | 15868 +++++++++++++++- .../tools/model_registry/hf_scraper.py | 98 +- .../tools/model_registry/relevancy.py | 135 + .../tools/model_registry/schemas.py | 17 +- 6 files changed, 19599 insertions(+), 1788 deletions(-) create mode 100644 tests/unit/tools/test_relevancy.py create mode 100644 transformer_lens/tools/model_registry/relevancy.py diff --git a/tests/unit/tools/test_relevancy.py b/tests/unit/tools/test_relevancy.py new file mode 100644 index 000000000..2cb2cd39e --- /dev/null +++ b/tests/unit/tools/test_relevancy.py @@ -0,0 +1,234 @@ +"""Unit tests for transformer_lens.tools.model_registry.relevancy. + +Tests cover: +- Individual component scoring functions +- Composite relevancy score computation +- Batch scoring and sorting via compute_scores_for_gaps +- Edge cases (zero values, missing data) +""" + +from transformer_lens.tools.model_registry.relevancy import ( + _normalize_demand, + _normalize_usage, + _score_benchmarkability, + compute_relevancy_score, + compute_scores_for_gaps, +) + +# ============================================================ +# _normalize_demand +# ============================================================ + + +class TestNormalizeDemand: + def test_max_count_returns_100(self): + assert _normalize_demand(50, 50) == 100.0 + + def test_half_of_max(self): + assert _normalize_demand(25, 50) == 50.0 + + def test_zero_count(self): + assert _normalize_demand(0, 50) == 0.0 + + def test_zero_max(self): + assert _normalize_demand(10, 0) == 0.0 + + def test_caps_at_100(self): + assert _normalize_demand(100, 50) == 100.0 + + +# ============================================================ +# _normalize_usage +# ============================================================ + + +class TestNormalizeUsage: + def test_max_downloads_returns_100(self): + assert _normalize_usage(1_000_000, 1_000_000) == 100.0 + + def test_zero_downloads(self): + assert _normalize_usage(0, 1_000_000) == 0.0 + + def test_zero_max(self): + assert _normalize_usage(100, 0) == 0.0 + + def test_log_scale_compresses_range(self): + # 10x fewer downloads should NOT produce 10x lower score (log compression) + score_high = _normalize_usage(1_000_000, 1_000_000) + score_low = _normalize_usage(100_000, 1_000_000) + assert score_low > score_high * 0.5 # Log scale keeps it above 50% of max + + def test_small_downloads_still_score(self): + score = _normalize_usage(100, 10_000_000) + assert score > 0 + + +# ============================================================ +# _score_benchmarkability +# ============================================================ + + +class TestScoreBenchmarkability: + def test_none_returns_zero(self): + assert _score_benchmarkability(None) == 0.0 + + def test_under_1b(self): + assert _score_benchmarkability(350_000_000) == 100.0 + + def test_exactly_1b(self): + assert _score_benchmarkability(1_000_000_000) == 100.0 + + def test_under_3b(self): + assert _score_benchmarkability(2_000_000_000) == 80.0 + + def test_under_7b(self): + assert _score_benchmarkability(6_000_000_000) == 60.0 + + def test_under_14b(self): + assert _score_benchmarkability(13_000_000_000) == 40.0 + + def test_under_30b(self): + assert _score_benchmarkability(27_000_000_000) == 20.0 + + def test_over_30b(self): + assert _score_benchmarkability(70_000_000_000) == 0.0 + + +# ============================================================ +# compute_relevancy_score +# ============================================================ + + +class TestComputeRelevancyScore: + def test_perfect_scores_all_components(self): + score = compute_relevancy_score( + model_count=100, + total_downloads=1_000_000, + min_param_count=350_000_000, + max_model_count=100, + max_downloads=1_000_000, + ) + assert score == 100.0 + + def test_zero_everything(self): + score = compute_relevancy_score( + model_count=0, + total_downloads=0, + min_param_count=None, + max_model_count=100, + max_downloads=1_000_000, + ) + assert score == 0.0 + + def test_score_in_valid_range(self): + score = compute_relevancy_score( + model_count=25, + total_downloads=500_000, + min_param_count=7_000_000_000, + max_model_count=100, + max_downloads=10_000_000, + ) + assert 0 <= score <= 100 + + def test_higher_downloads_increase_score(self): + base_kwargs = dict( + model_count=10, + min_param_count=1_000_000_000, + max_model_count=50, + max_downloads=10_000_000, + ) + score_low = compute_relevancy_score(total_downloads=1_000, **base_kwargs) + score_high = compute_relevancy_score(total_downloads=5_000_000, **base_kwargs) + assert score_high > score_low + + def test_smaller_model_increases_score(self): + base_kwargs = dict( + model_count=10, + total_downloads=100_000, + max_model_count=50, + max_downloads=10_000_000, + ) + score_big = compute_relevancy_score(min_param_count=70_000_000_000, **base_kwargs) + score_small = compute_relevancy_score(min_param_count=350_000_000, **base_kwargs) + assert score_small > score_big + + def test_higher_model_count_increases_score(self): + base_kwargs = dict( + total_downloads=100_000, + min_param_count=1_000_000_000, + max_model_count=100, + max_downloads=10_000_000, + ) + score_few = compute_relevancy_score(model_count=5, **base_kwargs) + score_many = compute_relevancy_score(model_count=80, **base_kwargs) + assert score_many > score_few + + +# ============================================================ +# compute_scores_for_gaps +# ============================================================ + + +class TestComputeScoresForGaps: + def test_adds_relevancy_score_field(self): + gaps = [ + { + "architecture_id": "CodeGenForCausalLM", + "total_models": 29, + "total_downloads": 100_000, + "min_param_count": 350_000_000, + }, + ] + result = compute_scores_for_gaps(gaps) + assert "relevancy_score" in result[0] + assert isinstance(result[0]["relevancy_score"], float) + + def test_sorts_by_score_descending(self): + gaps = [ + { + "architecture_id": "ArchLow", + "total_models": 2, + "total_downloads": 100, + "min_param_count": None, + }, + { + "architecture_id": "ArchHigh", + "total_models": 50, + "total_downloads": 4_000_000, + "min_param_count": 350_000_000, + }, + ] + result = compute_scores_for_gaps(gaps) + assert result[0]["architecture_id"] == "ArchHigh" + assert result[1]["architecture_id"] == "ArchLow" + assert result[0]["relevancy_score"] >= result[1]["relevancy_score"] + + def test_empty_list(self): + assert compute_scores_for_gaps([]) == [] + + def test_missing_optional_fields_use_defaults(self): + gaps = [ + {"architecture_id": "SomeNewArch", "total_models": 5}, + ] + result = compute_scores_for_gaps(gaps) + assert "relevancy_score" in result[0] + assert result[0]["relevancy_score"] >= 0 + + def test_tiebreaker_uses_model_count(self): + # Two architectures with same downloads and no params — demand decides + gaps = [ + { + "architecture_id": "ArchA", + "total_models": 10, + "total_downloads": 0, + "min_param_count": None, + }, + { + "architecture_id": "ArchB", + "total_models": 20, + "total_downloads": 0, + "min_param_count": None, + }, + ] + result = compute_scores_for_gaps(gaps) + assert result[0]["architecture_id"] == "ArchB" diff --git a/transformer_lens/tools/model_registry/data/architecture_gaps.json b/transformer_lens/tools/model_registry/data/architecture_gaps.json index 68ef2bda5..3728c5d98 100644 --- a/transformer_lens/tools/model_registry/data/architecture_gaps.json +++ b/transformer_lens/tools/model_registry/data/architecture_gaps.json @@ -1,17 +1,19 @@ { "generated_at": "2026-04-09", "scan_info": { - "total_scanned": 10000, + "total_scanned": 6354, "task_filter": "text-generation", "min_downloads": 500, - "scan_duration_seconds": 3.2 + "scan_duration_seconds": 12.1 }, - "total_unsupported_architectures": 372, - "total_unsupported_models": 1416, + "total_unsupported_architectures": 403, + "total_unsupported_models": 1594, "gaps": [ { "architecture_id": "Qwen3_5ForConditionalGeneration", - "total_models": 66, + "total_models": 76, + "total_downloads": 196957, + "min_param_count": 211968832, "sample_models": [ "Tesslate/OmniCoder-9B", "mconcat/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled-NVFP4", @@ -23,11 +25,14 @@ "ShinePixelOrg/Qwopus3.5-27B-v3-NVFP4", "aifeifei798/Qwen3.5-Queen-27B", "Jackrong/Qwen3.5-2B-Claude-4.6-Opus-Reasoning-Distilled" - ] + ], + "relevancy_score": 92.1 }, { "architecture_id": "Qwen3MoeForCausalLM", - "total_models": 55, + "total_models": 57, + "total_downloads": 5816931, + "min_param_count": 2574656, "sample_models": [ "Qwen/Qwen3-30B-A3B", "Qwen/Qwen3-30B-A3B-Instruct-2507", @@ -39,11 +44,14 @@ "Qwen/Qwen3-235B-A22B-Instruct-2507", "Qwen/Qwen3-Coder-480B-A35B-Instruct", "Qwen/Qwen3-235B-A22B-Thinking-2507" - ] + ], + "relevancy_score": 88.4 }, { "architecture_id": "DeepseekV3ForCausalLM", - "total_models": 51, + "total_models": 55, + "total_downloads": 6902421, + "min_param_count": 1656048, "sample_models": [ "deepseek-ai/DeepSeek-R1", "deepseek-ai/DeepSeek-R1-0528", @@ -55,11 +63,14 @@ "moonshotai/Kimi-K2-Instruct-0905", "moonshotai/Kimi-K2-Instruct", "moonshotai/Moonlight-16B-A3B-Instruct" - ] + ], + "relevancy_score": 87.6 }, { "architecture_id": "NemotronHForCausalLM", - "total_models": 50, + "total_models": 55, + "total_downloads": 5436161, + "min_param_count": 4221480, "sample_models": [ "nvidia/NVIDIA-Nemotron-3-Super-120B-A12B-NVFP4", "nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16", @@ -71,27 +82,14 @@ "nvidia/NVIDIA-Nemotron-3-Nano-4B-BF16", "unsloth/NVIDIA-Nemotron-3-Super-120B-A12B-NVFP4", "unsloth/NVIDIA-Nemotron-3-Nano-4B" - ] - }, - { - "architecture_id": "Lfm2ForCausalLM", - "total_models": 34, - "sample_models": [ - "farbodtavakkoli/OTel-LLM-1.2B-IT", - "LiquidAI/LFM2.5-1.2B-Instruct", - "LiquidAI/LFM2-1.2B", - "LiquidAI/LFM2-350M", - "LiquidAI/LFM2.5-1.2B-Thinking", - "LiquidAI/LFM2.5-350M", - "LiquidAI/LFM2-2.6B-Exp", - "LiquidAI/LFM2.5-1.2B-Base", - "LiquidAI/LFM2-700M", - "unsloth/LFM2.5-1.2B-Instruct" - ] + ], + "relevancy_score": 87.0 }, { "architecture_id": "Qwen3_5ForCausalLM", - "total_models": 32, + "total_models": 53, + "total_downloads": 78484, + "min_param_count": 752393024, "sample_models": [ "lukey03/Qwen3.5-9B-abliterated", "GoodStartLabs/gin-rummy-hbc-qwen3.5-0.8b", @@ -103,11 +101,33 @@ "kai-os/Carnice-9b", "aifeifei798/Darkidol-Ballad-9B", "continuum-ai/qwen3.5-4b-code-forged" - ] + ], + "relevancy_score": 76.4 + }, + { + "architecture_id": "Lfm2ForCausalLM", + "total_models": 39, + "total_downloads": 1371251, + "min_param_count": 274754048, + "sample_models": [ + "farbodtavakkoli/OTel-LLM-1.2B-IT", + "LiquidAI/LFM2.5-1.2B-Instruct", + "LiquidAI/LFM2-1.2B", + "LiquidAI/LFM2-350M", + "LiquidAI/LFM2.5-1.2B-Thinking", + "LiquidAI/LFM2.5-350M", + "LiquidAI/LFM2-2.6B-Exp", + "LiquidAI/LFM2.5-1.2B-Base", + "LiquidAI/LFM2-700M", + "unsloth/LFM2.5-1.2B-Instruct" + ], + "relevancy_score": 74.5 }, { "architecture_id": "Gemma4ForConditionalGeneration", - "total_models": 30, + "total_models": 37, + "total_downloads": 514990, + "min_param_count": 1198399786, "sample_models": [ "nvidia/Gemma-4-31B-IT-NVFP4", "dealignai/Gemma-4-31B-JANG_4M-CRACK", @@ -118,76 +138,15 @@ "bg-digitalservices/Gemma-4-26B-A4B-it-NVFP4A16", "0xSero/gemma-4-21b-a4b-it-REAP", "InfinimindCreations/gemma-4-E4B-it-uncensored", - "EganAI/gemma-4-31B-Claude-4.6-Opus-Reasoning-Distilled" - ] - }, - { - "architecture_id": "CodeGenForCausalLM", - "total_models": 29, - "sample_models": [ - "Salesforce/codegen-350M-mono", - "Salesforce/codegen-350M-multi", - "Salesforce/codegen-2B-mono", - "Salesforce/codegen-6B-multi", - "Salesforce/codegen-16B-nl", - "Salesforce/codegen-6B-nl", - "Salesforce/codegen-350M-nl", - "Salesforce/codegen-6B-mono", - "Salesforce/codegen-2B-multi", - "Salesforce/codegen-16B-mono" - ] - }, - { - "architecture_id": "MPTForCausalLM", - "total_models": 24, - "sample_models": [ - "vinai/PhoGPT-4B", - "anas-awadalla/mpt-7b", - "gl198976/mpt-7b-instruct", - "replit/replit-code-v1-3b", - "vinai/PhoGPT-4B-Chat", - "wtang06/mpt-125m-c4", - "echarlaix/tiny-mpt-random-remote-code", - "lightblue/japanese-mpt-7b", - "gl198976/mpt-7b", - "TehVenom/MPT-7b-InstructAndStorywriting-50_50-Merge" - ] - }, - { - "architecture_id": "Qwen3_5MoeForConditionalGeneration", - "total_models": 23, - "sample_models": [ - "nvidia/Qwen3.5-397B-A17B-NVFP4", - "txn545/Qwen3.5-122B-A10B-NVFP4", - "Jackrong/Qwen3.5-35B-A3B-Claude-4.6-Opus-Reasoning-Distilled", - "lukealonso/Qwen3.5-397B-A17B-NVFP4", - "txn545/Qwen3.5-35B-A3B-NVFP4", - "nightmedia/Qwen3.5-122B-A10B-Text-mxfp4-mlx", - "olka-fi/Qwen3.5-122B-A10B-MXFP4", - "nightmedia/Qwen3.5-35B-A3B-Text-qx64-hi-mlx", - "RepublicOfKorokke/Qwen3.5-35B-A3B-mlx-lm-mxfp4", - "bjk110/Qwen3.5-122B-A10B-abliterated-NVFP4" - ] - }, - { - "architecture_id": "InternLM2ForCausalLM", - "total_models": 23, - "sample_models": [ - "internlm/internlm2-chat-7b", - "internlm/internlm2_5-7b-chat", - "internlm/internlm2-7b", - "internlm/internlm2-20b", - "internlm/internlm2-base-7b", - "internlm/internlm2-chat-20b", - "internlm/internlm2-base-20b", - "chujiezheng/internlm2-chat-20b-ExPO", - "chujiezheng/internlm2-chat-7b-ExPO", - "internlm/internlm2-1_8b" - ] + "livadies/gemma-4-E2B-Ghetto-NF4" + ], + "relevancy_score": 67.1 }, { "architecture_id": "Qwen3NextForCausalLM", - "total_models": 21, + "total_models": 25, + "total_downloads": 1770629, + "min_param_count": 2839160, "sample_models": [ "Qwen/Qwen3-Coder-Next", "Qwen/Qwen3-Next-80B-A3B-Instruct", @@ -199,27 +158,33 @@ "RedHatAI/Qwen3-Coder-Next-NVFP4", "yujiepan/qwen3-next-moe-tiny-random", "saricles/Qwen3-Coder-Next-NVFP4-GB10" - ] + ], + "relevancy_score": 66.8 }, { - "architecture_id": "JambaForCausalLM", - "total_models": 21, + "architecture_id": "CodeGenForCausalLM", + "total_models": 30, + "total_downloads": 228380, + "min_param_count": 3877376, "sample_models": [ - "ai21labs/AI21-Jamba-Mini-1.5", - "ai21labs/Jamba-tiny-random", - "ai21labs/AI21-Jamba-Mini-1.6", - "ai21labs/AI21-Jamba-Large-1.5", - "ai21labs/AI21-Jamba2-3B", - "ai21labs/AI21-Jamba-Large-1.6", - "ai21labs/Jamba-v0.1", - "ai21labs/AI21-Jamba2-Mini", - "ai21labs/AI21-Jamba-Reasoning-3B", - "microsoft/Dayhoff-170m-GR" - ] + "Salesforce/codegen-350M-mono", + "Salesforce/codegen-350M-multi", + "Salesforce/codegen-2B-mono", + "Salesforce/codegen-6B-multi", + "Salesforce/codegen-16B-nl", + "Salesforce/codegen-6B-nl", + "Salesforce/codegen-350M-nl", + "Salesforce/codegen-6B-mono", + "Salesforce/codegen-2B-multi", + "Salesforce/codegen-16B-mono" + ], + "relevancy_score": 65.2 }, { "architecture_id": "QWenLMHeadModel", - "total_models": 20, + "total_models": 22, + "total_downloads": 505425, + "min_param_count": 19545408, "sample_models": [ "cckevinn/SeeClick", "Qwen/Qwen-7B-Chat", @@ -231,11 +196,33 @@ "Qwen/Qwen-14B", "Xingyu-Zheng/Qwen-VL-Chat", "Qwen/Qwen-72B" - ] + ], + "relevancy_score": 62.2 + }, + { + "architecture_id": "InternLM2ForCausalLM", + "total_models": 24, + "total_downloads": 257695, + "min_param_count": 24052864, + "sample_models": [ + "internlm/internlm2-chat-7b", + "internlm/internlm2_5-7b-chat", + "internlm/internlm2-7b", + "internlm/internlm2-20b", + "internlm/internlm2-base-7b", + "internlm/internlm2-chat-20b", + "internlm/internlm2-base-20b", + "chujiezheng/internlm2-chat-20b-ExPO", + "chujiezheng/internlm2-chat-7b-ExPO", + "internlm/internlm2-1_8b" + ], + "relevancy_score": 61.9 }, { "architecture_id": "GPTBigCodeForCausalLM", - "total_models": 20, + "total_models": 25, + "total_downloads": 116213, + "min_param_count": 1845928, "sample_models": [ "bigcode/gpt_bigcode-santacoder", "bigcode/tiny_starcoder_py", @@ -247,11 +234,52 @@ "defog/sqlcoder2", "HuggingFaceH4/starchat-beta", "LoupGarou/WizardCoder-Guanaco-15B-V1.0" - ] + ], + "relevancy_score": 60.7 + }, + { + "architecture_id": "Glm4MoeForCausalLM", + "total_models": 15, + "total_downloads": 755226, + "min_param_count": 2572352, + "sample_models": [ + "zai-org/GLM-4.5-Air", + "zai-org/GLM-4.7", + "trl-internal-testing/tiny-Glm4MoeForCausalLM", + "zai-org/GLM-4.5", + "zai-org/GLM-4.6", + "Tengyunw/GLM-4.7-NVFP4", + "np-cr/testing-glm4-moe", + "nvidia/GLM-4.7-NVFP4", + "Salyut1/GLM-4.7-NVFP4", + "ArliAI/GLM-4.6-Derestricted-v3" + ], + "relevancy_score": 59.0 + }, + { + "architecture_id": "T5GemmaForConditionalGeneration", + "total_models": 13, + "total_downloads": 1014188, + "min_param_count": 312517632, + "sample_models": [ + "google/t5gemma-s-s-prefixlm", + "google/t5gemma-9b-9b-ul2", + "google/t5gemma-b-b-ul2", + "google/t5gemma-2b-2b-prefixlm", + "google/t5gemma-2b-2b-ul2", + "google/t5gemma-l-l-ul2-it", + "google/t5gemma-ml-ml-ul2-it", + "google/t5gemma-b-b-prefixlm", + "google/t5gemma-s-s-prefixlm-it", + "google/t5gemma-s-s-ul2" + ], + "relevancy_score": 58.4 }, { "architecture_id": "XGLMForCausalLM", "total_models": 18, + "total_downloads": 223134, + "min_param_count": 162256896, "sample_models": [ "facebook/xglm-564M", "facebook/incoder-1B", @@ -263,43 +291,123 @@ "KoboldAI/fairseq-dense-355M", "KoboldAI/fairseq-dense-13B", "KoboldAI/fairseq-dense-1.3B" - ] + ], + "relevancy_score": 58.0 }, { - "architecture_id": "MiniMaxM2ForCausalLM", - "total_models": 14, + "architecture_id": "Qwen3_5MoeForConditionalGeneration", + "total_models": 27, + "total_downloads": 681338, + "min_param_count": 6643527536, "sample_models": [ - "MiniMaxAI/MiniMax-M2.5", - "cerebras/MiniMax-M2.1-REAP-139B-A10B", - "MiniMaxAI/MiniMax-M2", - "MiniMaxAI/MiniMax-M2.1", - "nvidia/MiniMax-M2.5-NVFP4", - "cerebras/MiniMax-M2.5-REAP-139B-A10B", - "amd/MiniMax-M2.5-MXFP4", - "saricles/MiniMax-M2.5-REAP-172B-A10B-NVFP4-GB10", - "aspctu/MiniMax-M2.5", - "amd/MiniMax-M2.1-MXFP4" - ] + "nvidia/Qwen3.5-397B-A17B-NVFP4", + "txn545/Qwen3.5-122B-A10B-NVFP4", + "Jackrong/Qwen3.5-35B-A3B-Claude-4.6-Opus-Reasoning-Distilled", + "lukealonso/Qwen3.5-397B-A17B-NVFP4", + "txn545/Qwen3.5-35B-A3B-NVFP4", + "nightmedia/Qwen3.5-122B-A10B-Text-mxfp4-mlx", + "olka-fi/Qwen3.5-122B-A10B-MXFP4", + "nightmedia/Qwen3.5-35B-A3B-Text-qx64-hi-mlx", + "RepublicOfKorokke/Qwen3.5-35B-A3B-mlx-lm-mxfp4", + "bjk110/Qwen3.5-122B-A10B-abliterated-NVFP4" + ], + "relevancy_score": 57.8 }, { - "architecture_id": "DeciLMForCausalLM", + "architecture_id": "JambaForCausalLM", + "total_models": 22, + "total_downloads": 43853, + "min_param_count": 127679344, + "sample_models": [ + "ai21labs/AI21-Jamba-Mini-1.5", + "ai21labs/Jamba-tiny-random", + "ai21labs/AI21-Jamba-Mini-1.6", + "ai21labs/AI21-Jamba-Large-1.5", + "ai21labs/AI21-Jamba2-3B", + "ai21labs/AI21-Jamba-Large-1.6", + "ai21labs/Jamba-v0.1", + "ai21labs/AI21-Jamba2-Mini", + "ai21labs/AI21-Jamba-Reasoning-3B", + "microsoft/Dayhoff-170m-GR" + ], + "relevancy_score": 56.8 + }, + { + "architecture_id": "DeepseekV32ForCausalLM", + "total_models": 9, + "total_downloads": 1416598, + "min_param_count": 136559748, + "sample_models": [ + "deepseek-ai/DeepSeek-V3.2", + "deepseek-ai/DeepSeek-V3.2-Exp", + "nvidia/DeepSeek-V3.2-NVFP4", + "deepseek-ai/DeepSeek-V3.2-Speciale", + "deepseek-ai/DeepSeek-Math-V2", + "exolabs/DeepSeek-V3.2_bf16", + "deepseek-ai/DeepSeek-V3.2-Exp-Base", + "hyper-accel/tiny-random-deepseek-v32", + "cs2764/DeepSeek-V3.2_dq4-mlx" + ], + "relevancy_score": 56.8 + }, + { + "architecture_id": "SmolLM3ForCausalLM", + "total_models": 7, + "total_downloads": 1166052, + "min_param_count": 8245568, + "sample_models": [ + "HuggingFaceTB/SmolLM3-3B", + "HuggingFaceTB/SmolLM3-3B-Base", + "optimum-internal-testing/tiny-random-SmolLM3ForCausalLM", + "unsloth/SmolLM3-3B", + "onnx-internal-testing/tiny-random-SmolLM3ForCausalLM", + "MInAlA/smollm3-dpo-merged", + "N-Bot-Int/SmolSam3-MEMGRPO" + ], + "relevancy_score": 55.2 + }, + { + "architecture_id": "BartForConditionalGeneration", + "total_models": 9, + "total_downloads": 567852, + "min_param_count": 6044480, + "sample_models": [ + "KomeijiForce/bart-large-emojilm", + "antalvdb/bart-base-spelling-nl", + "lmqg/bart-large-squad-qg", + "kengurukleo/deutsch_a2_transformer", + "shibing624/bart4csc-base-chinese", + "SkitCon/gec-spanish-BARTO-SYNTHETIC", + "Nargizi/screeve-lemmatizer", + "Tianlin668/MentalBART", + "KomeijiForce/bart-large-emojilm-e2t" + ], + "relevancy_score": 54.8 + }, + { + "architecture_id": "BaichuanForCausalLM", "total_models": 14, + "total_downloads": 121942, + "min_param_count": 16204352, "sample_models": [ - "nvidia/Llama-3_3-Nemotron-Super-49B-v1_5", - "nvidia/Llama-3_3-Nemotron-Super-49B-v1", - "nvidia/Llama-3_3-Nemotron-Super-49B-v1_5-NVFP4", - "nvidia/Llama-3_1-Nemotron-Ultra-253B-v1", - "ConicCat/Llama3_3-Nemo-Super-Writer-49B", - "nvidia/Llama-3_1-Nemotron-51B-Instruct", - "FriendliAI/Llama-3_3-Nemotron-Super-49B-v1_5", - "FriendliAI/Llama-3_1-Nemotron-Ultra-253B-v1", - "NewstaR/Porpoise-6b-instruct", - "nvidia/Llama-3_1-Nemotron-Ultra-253B-CPT-v1" - ] + "baichuan-inc/Baichuan2-7B-Chat", + "baichuan-inc/Baichuan2-13B-Chat", + "baichuan-inc/Baichuan-13B-Chat", + "baichuan-inc/Baichuan2-7B-Base", + "baichuan-inc/Baichuan2-13B-Base", + "katuni4ka/tiny-random-baichuan2", + "sakuraumi/Sakura-13B-Galgame", + "zxbsmk/NSFW_13B_sft", + "katuni4ka/tiny-random-baichuan2-13b", + "baichuan-inc/Baichuan-13B-Base" + ], + "relevancy_score": 54.3 }, { "architecture_id": "FalconH1ForCausalLM", - "total_models": 14, + "total_models": 15, + "total_downloads": 76716, + "min_param_count": 91131072, "sample_models": [ "tiiuae/Falcon-H1-0.5B-Base", "tiiuae/Falcon-H1-3B-Base", @@ -311,107 +419,127 @@ "tiiuae/Falcon-H1-Tiny-90M-Instruct", "tiiuae/Falcon-H1-1.5B-Deep-Instruct", "tiiuae/Falcon-H1-3B-Instruct" - ] + ], + "relevancy_score": 53.9 }, { - "architecture_id": "RwkvForCausalLM", - "total_models": 14, + "architecture_id": "H2OVLChatModel", + "total_models": 2, + "total_downloads": 2156965, + "min_param_count": 826295808, "sample_models": [ - "RWKV/v5-Eagle-7B-HF", - "RWKV/rwkv-4-169m-pile", - "beomi/KoRWKV-6B", - "RWKV/rwkv-4-430m-pile", - "RWKV/rwkv-4-1b5-pile", - "RWKV/rwkv-4-3b-pile", - "RWKV/rwkv-raven-1b5", - "RWKV/rwkv-4-7b-pile", - "RWKV/rwkv-raven-3b", - "RWKV/rwkv-raven-14b" - ] + "h2oai/h2ovl-mississippi-800m", + "h2oai/h2ovl-mississippi-2b" + ], + "relevancy_score": 53.6 }, { - "architecture_id": "DeepseekV2ForCausalLM", - "total_models": 13, + "architecture_id": "GlmMoeDsaForCausalLM", + "total_models": 7, + "total_downloads": 529214, + "min_param_count": 162774148, "sample_models": [ - "deepseek-ai/DeepSeek-V2-Lite-Chat", - "deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct", - "deepseek-ai/DeepSeek-V2-Lite", - "deepseek-ai/DeepSeek-V2", - "deepseek-ai/DeepSeek-V2-Chat", - "deepseek-ai/DeepSeek-Coder-V2-Instruct", - "deepseek-ai/DeepSeek-V2.5", - "deepseek-ai/DeepSeek-V2-Chat-0628", - "deepseek-ai/DeepSeek-Coder-V2-Lite-Base", - "Kwaipilot/KwaiCoder-DS-V2-Lite-Base" - ] + "zai-org/GLM-5", + "nvidia/GLM-5-NVFP4", + "zai-org/GLM-5.1", + "cs2764/GLM-5-abliterated-dq4-mlx", + "0xSero/GLM-5-REAP-381B", + "cs2764/GLM-5-abliterated-dq3-mlx", + "hyper-accel/tiny-random-glm-moe-dsa" + ], + "relevancy_score": 53.4 }, { - "architecture_id": "Glm4MoeForCausalLM", - "total_models": 13, + "architecture_id": "CohereForCausalLM", + "total_models": 10, + "total_downloads": 184596, + "min_param_count": 2042176, "sample_models": [ - "zai-org/GLM-4.5-Air", - "zai-org/GLM-4.7", - "trl-internal-testing/tiny-Glm4MoeForCausalLM", - "zai-org/GLM-4.5", - "zai-org/GLM-4.6", - "Tengyunw/GLM-4.7-NVFP4", - "np-cr/testing-glm4-moe", - "nvidia/GLM-4.7-NVFP4", - "Salyut1/GLM-4.7-NVFP4", - "ArliAI/GLM-4.6-Derestricted-v3" - ] + "trl-internal-testing/tiny-CohereForCausalLM", + "CohereLabs/aya-expanse-8b", + "CohereLabs/c4ai-command-r-v01", + "CohereLabs/aya-23-8B", + "NLPark/AnFeng_v3_Avocet", + "CohereLabs/aya-expanse-32b", + "CohereLabs/aya-23-35B", + "CohereLabs/c4ai-command-r-plus-08-2024", + "CohereLabs/c4ai-command-r-08-2024", + "CohereLabs/c4ai-command-r-plus" + ], + "relevancy_score": 52.9 }, { - "architecture_id": "BaichuanForCausalLM", - "total_models": 13, + "architecture_id": "PhiMoEForCausalLM", + "total_models": 4, + "total_downloads": 870913, + "min_param_count": 1110112, "sample_models": [ - "baichuan-inc/Baichuan2-7B-Chat", - "baichuan-inc/Baichuan2-13B-Chat", - "baichuan-inc/Baichuan-13B-Chat", - "baichuan-inc/Baichuan2-7B-Base", - "baichuan-inc/Baichuan2-13B-Base", - "katuni4ka/tiny-random-baichuan2", - "sakuraumi/Sakura-13B-Galgame", - "zxbsmk/NSFW_13B_sft", - "katuni4ka/tiny-random-baichuan2-13b", - "baichuan-inc/Baichuan-13B-Base" - ] + "microsoft/Phi-tiny-MoE-instruct", + "microsoft/Phi-mini-MoE-instruct", + "microsoft/Phi-3.5-MoE-instruct", + "optimum-intel-internal-testing/phi-3.5-moe-tiny-random" + ], + "relevancy_score": 52.8 }, { - "architecture_id": "LlavaLlamaForCausalLM", - "total_models": 13, + "architecture_id": "MiniCPMForCausalLM", + "total_models": 12, + "total_downloads": 94018, + "min_param_count": 80000640, "sample_models": [ - "LanguageBind/Video-LLaVA-7B", - "wisdomik/Quilt-Llava-v1.5-7b", - "liuhaotian/llava-llama-2-13b-chat-lightning-preview", - "lmms-lab/llama3-llava-next-8b", - "mmaaz60/LLaVA-7B-Lightening-v1-1", - "microsoft/llava-med-7b-delta", - "deepcs233/VisCoT-7b-336", - "ManishThota/Ollama_Video_llama_7B", - "EricPolaris/Quilt-Llava-v1.5-7b", - "liuhaotian/LLaVA-Lightning-7B-delta-v1-1" - ] + "openbmb/MiniCPM-2B-sft-bf16", + "openbmb/MiniCPM4.1-8B", + "openbmb/MiniCPM-1B-sft-bf16", + "openbmb/MiniCPM4-0.5B", + "openbmb/MiniCPM-MoE-8x2B", + "katuni4ka/tiny-random-minicpm", + "openbmb/MiniCPM-S-1B-sft", + "openbmb/MiniCPM-2B-sft-fp32", + "openbmb/MiniCPM-2B-dpo-bf16", + "openbmb/MiniCPM4-8B" + ], + "relevancy_score": 52.6 }, { - "architecture_id": "T5GemmaForConditionalGeneration", - "total_models": 12, + "architecture_id": "Cohere2ForCausalLM", + "total_models": 9, + "total_downloads": 199383, + "min_param_count": 2090024, "sample_models": [ - "google/t5gemma-s-s-prefixlm", - "google/t5gemma-9b-9b-ul2", - "google/t5gemma-b-b-ul2", - "google/t5gemma-2b-2b-prefixlm", - "google/t5gemma-2b-2b-ul2", - "google/t5gemma-l-l-ul2-it", - "google/t5gemma-ml-ml-ul2-it", - "google/t5gemma-b-b-prefixlm", - "google/t5gemma-s-s-prefixlm-it", - "google/t5gemma-s-s-ul2" - ] + "trl-internal-testing/tiny-Cohere2ForCausalLM", + "CohereLabs/tiny-aya-global", + "CohereLabs/c4ai-command-r7b-12-2024", + "CohereLabs/tiny-aya-base", + "CohereLabs/c4ai-command-r7b-arabic-02-2025", + "CohereLabs/c4ai-command-a-03-2025", + "CohereLabs/tiny-aya-water", + "CohereLabs/tiny-aya-fire", + "CohereLabs/tiny-aya-earth" + ], + "relevancy_score": 52.5 + }, + { + "architecture_id": "MambaForCausalLM", + "total_models": 8, + "total_downloads": 253701, + "min_param_count": 129135360, + "sample_models": [ + "state-spaces/mamba-130m-hf", + "state-spaces/mamba-2.8b-hf", + "state-spaces/mamba-1.4b-hf", + "state-spaces/mamba-370m-hf", + "state-spaces/mamba-790m-hf", + "NYTK/PULI-HuBA-mamba-130M", + "EchoLabs33/mamba-130m-hxq", + "TRI-ML/mamba-7b-rw" + ], + "relevancy_score": 52.4 }, { "architecture_id": "MT5ForConditionalGeneration", - "total_models": 12, + "total_models": 13, + "total_downloads": 56841, + "min_param_count": 300176768, "sample_models": [ "knowledgator/IUPAC2SMILES-canonical-base", "knowledgator/SMILES2IUPAC-canonical-base", @@ -423,43 +551,87 @@ "intelia-lab-uah/mt0-base_QG_SQAC", "intelia-lab-uah/mt0-base_AE_SQAC", "UBC-NLP/toucan-1.2B" - ] + ], + "relevancy_score": 52.0 }, { - "architecture_id": "LLaMAForCausalLM", - "total_models": 12, + "architecture_id": "RwkvForCausalLM", + "total_models": 15, + "total_downloads": 31600, + "min_param_count": 169342464, "sample_models": [ - "maicomputer/alpaca-13b", - "Enoch/llama-65b-hf", - "mncai/chatdoctor", - "AdaptLLM/law-LLM", - "Nitish-Garikoti/finance-LLM", - "boboto/LLaMA-65B-HF", - "AdaptLLM/finance-LLM", - "AdaptLLM/medicine-LLM", - "Rardilit/Panther_v1", - "James-WYang/BigTranslate" - ] + "RWKV/v5-Eagle-7B-HF", + "RWKV/rwkv-4-169m-pile", + "beomi/KoRWKV-6B", + "RWKV/rwkv-4-430m-pile", + "RWKV/rwkv-4-1b5-pile", + "RWKV/rwkv-4-3b-pile", + "RWKV/rwkv-raven-1b5", + "RWKV/rwkv-4-7b-pile", + "RWKV/rwkv-raven-3b", + "RWKV/rwkv-raven-14b" + ], + "relevancy_score": 51.9 }, { - "architecture_id": "MiniCPMForCausalLM", - "total_models": 11, + "architecture_id": "DFlashDraftModel", + "total_models": 10, + "total_downloads": 121528, + "min_param_count": 473995264, "sample_models": [ - "openbmb/MiniCPM-2B-sft-bf16", - "openbmb/MiniCPM4.1-8B", - "openbmb/MiniCPM-1B-sft-bf16", - "openbmb/MiniCPM4-0.5B", - "openbmb/MiniCPM-MoE-8x2B", - "katuni4ka/tiny-random-minicpm", - "openbmb/MiniCPM-S-1B-sft", - "openbmb/MiniCPM-2B-sft-fp32", - "openbmb/MiniCPM-2B-dpo-bf16", - "openbmb/MiniCPM4-8B" - ] + "z-lab/Qwen3-4B-DFlash-b16", + "z-lab/Qwen3-8B-DFlash-b16", + "z-lab/Qwen3.5-9B-DFlash", + "z-lab/Qwen3.5-4B-DFlash", + "z-lab/Qwen3.5-27B-DFlash", + "z-lab/gpt-oss-20b-DFlash", + "z-lab/gpt-oss-120b-DFlash", + "z-lab/Qwen3.5-35B-A3B-DFlash", + "z-lab/LLaMA3.1-8B-Instruct-DFlash-UltraChat", + "z-lab/Qwen3-Coder-30B-A3B-DFlash" + ], + "relevancy_score": 51.9 + }, + { + "architecture_id": "Qwen2MoeForCausalLM", + "total_models": 7, + "total_downloads": 215317, + "min_param_count": 1219036, + "sample_models": [ + "Qwen/Qwen1.5-MoE-A2.7B", + "Qwen/Qwen1.5-MoE-A2.7B-Chat", + "Qwen/Qwen2-57B-A14B-Instruct", + "Qwen/Qwen2-57B-A14B", + "katuni4ka/tiny-random-qwen1.5-moe", + "yujiepan/qwen1.5-moe-tiny-random", + "xd2010/Qwen1.5-MOE-sft-math7k-densemixer" + ], + "relevancy_score": 51.4 + }, + { + "architecture_id": "MPTForCausalLM", + "total_models": 26, + "total_downloads": 36294, + "min_param_count": 6649286656, + "sample_models": [ + "vinai/PhoGPT-4B", + "anas-awadalla/mpt-7b", + "gl198976/mpt-7b-instruct", + "replit/replit-code-v1-3b", + "vinai/PhoGPT-4B-Chat", + "wtang06/mpt-125m-c4", + "echarlaix/tiny-mpt-random-remote-code", + "lightblue/japanese-mpt-7b", + "gl198976/mpt-7b", + "TehVenom/MPT-7b-InstructAndStorywriting-50_50-Merge" + ], + "relevancy_score": 50.7 }, { "architecture_id": "Glm4MoeLiteForCausalLM", - "total_models": 10, + "total_models": 12, + "total_downloads": 1342614, + "min_param_count": 4866195072, "sample_models": [ "zai-org/GLM-4.7-Flash", "unsloth/GLM-4.7-Flash", @@ -471,89 +643,145 @@ "jerrycheng233/model5_sft_16bit", "aaravriyer193/chimpgpt-coder-elite", "GadflyII/GLM-4.7-Flash-MTP-NVFP4" - ] + ], + "relevancy_score": 50.5 }, { - "architecture_id": "CohereForCausalLM", - "total_models": 10, + "architecture_id": "Phi3VForCausalLM", + "total_models": 6, + "total_downloads": 174687, + "min_param_count": 304612720, "sample_models": [ - "trl-internal-testing/tiny-CohereForCausalLM", - "CohereLabs/aya-expanse-8b", - "CohereLabs/c4ai-command-r-v01", - "CohereLabs/aya-23-8B", - "NLPark/AnFeng_v3_Avocet", - "CohereLabs/aya-expanse-32b", - "CohereLabs/aya-23-35B", - "CohereLabs/c4ai-command-r-plus-08-2024", - "CohereLabs/c4ai-command-r-08-2024", - "CohereLabs/c4ai-command-r-plus" - ] + "microsoft/Phi-3-vision-128k-instruct", + "TIGER-Lab/VLM2Vec-Full", + "yujiepan/phi-3-vision-tiny-random", + "furonghuang-lab/tracevla_phi3v", + "Desm0nt/Phi-3-HornyVision-128k-instruct", + "failspy/Phi-3-vision-128k-instruct-abliterated-alpha" + ], + "relevancy_score": 50.4 }, { - "architecture_id": "DFlashDraftModel", - "total_models": 10, + "architecture_id": "FalconMambaForCausalLM", + "total_models": 5, + "total_downloads": 183307, + "min_param_count": 525400, "sample_models": [ - "z-lab/Qwen3-4B-DFlash-b16", - "z-lab/Qwen3-8B-DFlash-b16", - "z-lab/Qwen3.5-9B-DFlash", - "z-lab/Qwen3.5-4B-DFlash", - "z-lab/Qwen3.5-27B-DFlash", - "z-lab/gpt-oss-20b-DFlash", - "z-lab/gpt-oss-120b-DFlash", - "z-lab/Qwen3.5-35B-A3B-DFlash", - "z-lab/LLaMA3.1-8B-Instruct-DFlash-UltraChat", - "z-lab/Qwen3-Coder-30B-A3B-DFlash" - ] + "trl-internal-testing/tiny-FalconMambaForCausalLM", + "tiiuae/falcon-mamba-7b-instruct", + "tiiuae/falcon-mamba-7b", + "tiiuae/falcon-mamba-tiny-dev", + "tiiuae/Falcon3-Mamba-7B-Instruct" + ], + "relevancy_score": 49.9 }, { - "architecture_id": "RWForCausalLM", - "total_models": 10, + "architecture_id": "ExaoneForCausalLM", + "total_models": 7, + "total_downloads": 620915, + "min_param_count": 2405327360, "sample_models": [ - "projecte-aina/aguila-7b", - "lightonai/alfred-40b-1023", - "explosion-testing/refined-web-model-test", - "vilm/vulture-40b", - "h2oai/h2ogpt-gm-oasst1-en-2048-falcon-7b-v2", - "nomic-ai/gpt4all-falcon", - "h2oai/h2ogpt-gm-oasst1-en-2048-falcon-7b-v3", - "OpenAssistant/falcon-40b-sft-top1-560", - "QuixiAI/WizardLM-Uncensored-Falcon-40b", - "mrm8488/falcoder-7b" - ] + "LGAI-EXAONE/EXAONE-Deep-7.8B", + "LGAI-EXAONE/EXAONE-3.5-7.8B-Instruct", + "LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct", + "LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct", + "LGAI-EXAONE/EXAONE-3.5-32B-Instruct", + "LGAI-EXAONE/EXAONE-Deep-32B", + "LGAI-EXAONE/EXAONE-Deep-2.4B" + ], + "relevancy_score": 49.8 }, { - "architecture_id": "DeepseekV32ForCausalLM", - "total_models": 9, + "architecture_id": "BambaForCausalLM", + "total_models": 3, + "total_downloads": 224385, + "min_param_count": 33110760, "sample_models": [ - "deepseek-ai/DeepSeek-V3.2", - "deepseek-ai/DeepSeek-V3.2-Exp", - "nvidia/DeepSeek-V3.2-NVFP4", - "deepseek-ai/DeepSeek-V3.2-Speciale", - "deepseek-ai/DeepSeek-Math-V2", - "exolabs/DeepSeek-V3.2_bf16", - "deepseek-ai/DeepSeek-V3.2-Exp-Base", - "hyper-accel/tiny-random-deepseek-v32", - "cs2764/DeepSeek-V3.2_dq4-mlx" - ] + "hmellor/tiny-random-BambaForCausalLM", + "ibm-ai-platform/Bamba-9B-v1", + "ibm-ai-platform/Bamba-9B-v2" + ], + "relevancy_score": 49.2 }, { - "architecture_id": "Cohere2ForCausalLM", - "total_models": 9, + "architecture_id": "LlavaQwenForCausalLM", + "total_models": 4, + "total_downloads": 162169, + "min_param_count": 893618208, "sample_models": [ - "trl-internal-testing/tiny-Cohere2ForCausalLM", - "CohereLabs/tiny-aya-global", - "CohereLabs/c4ai-command-r7b-12-2024", - "CohereLabs/tiny-aya-base", - "CohereLabs/c4ai-command-r7b-arabic-02-2025", - "CohereLabs/c4ai-command-a-03-2025", - "CohereLabs/tiny-aya-water", - "CohereLabs/tiny-aya-fire", - "CohereLabs/tiny-aya-earth" - ] + "lmms-lab/llava-onevision-qwen2-7b-ov", + "lmms-lab/llava-onevision-qwen2-0.5b-ov", + "lmms-lab/llava-onevision-qwen2-0.5b-si", + "lmms-lab/llava-onevision-qwen2-7b-si" + ], + "relevancy_score": 49.0 + }, + { + "architecture_id": "Eagle3Speculator", + "total_models": 5, + "total_downloads": 106067, + "min_param_count": 950186496, + "sample_models": [ + "RedHatAI/Qwen3-8B-speculator.eagle3", + "RedHatAI/Llama-3.1-8B-Instruct-speculator.eagle3", + "RedHatAI/Llama-3.3-70B-Instruct-speculator.eagle3", + "RedHatAI/Qwen3-32B-speculator.eagle3", + "RedHatAI/Qwen3-14B-speculator.eagle3" + ], + "relevancy_score": 48.7 + }, + { + "architecture_id": "OpenAIGPTLMHeadModel", + "total_models": 2, + "total_downloads": 234004, + "min_param_count": 119680512, + "sample_models": [ + "openai-community/openai-gpt", + "lgaalves/gpt1" + ], + "relevancy_score": 48.7 + }, + { + "architecture_id": "DeepseekV2ForCausalLM", + "total_models": 15, + "total_downloads": 1536809, + "min_param_count": 9019954810, + "sample_models": [ + "deepseek-ai/DeepSeek-V2-Lite-Chat", + "deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct", + "deepseek-ai/DeepSeek-V2-Lite", + "deepseek-ai/DeepSeek-V2", + "deepseek-ai/DeepSeek-V2-Chat", + "deepseek-ai/DeepSeek-Coder-V2-Instruct", + "deepseek-ai/DeepSeek-V2.5", + "deepseek-ai/DeepSeek-V2-Chat-0628", + "deepseek-ai/DeepSeek-Coder-V2-Lite-Base", + "Kwaipilot/KwaiCoder-DS-V2-Lite-Base" + ], + "relevancy_score": 48.5 + }, + { + "architecture_id": "BloomModel", + "total_models": 8, + "total_downloads": 38163, + "min_param_count": 16156544, + "sample_models": [ + "bigscience/bigscience-small-testing", + "TurkuNLP/gpt3-finnish-small", + "TurkuNLP/gpt3-finnish-large", + "TurkuNLP/gpt3-finnish-13B", + "BelleGroup/BELLE-7B-2M", + "norallm/norbloom-7b-scratch", + "Muennighoff/bloom-tiny-random", + "TurkuNLP/gpt3-finnish-xl" + ], + "relevancy_score": 48.2 }, { "architecture_id": "HunYuanDenseV1ForCausalLM", "total_models": 9, + "total_downloads": 28426, + "min_param_count": 539010048, "sample_models": [ "tencent/Hunyuan-7B-Instruct", "tencent/Hunyuan-0.5B-Pretrain", @@ -564,95 +792,123 @@ "tencent/Hunyuan-1.8B-Instruct", "tencent/Hunyuan-0.5B-Instruct", "tencent/Hunyuan-4B-Instruct" - ] + ], + "relevancy_score": 48.1 }, { - "architecture_id": "HybridQwen3ForCausalLM", - "total_models": 9, + "architecture_id": "ProGenForCausalLM", + "total_models": 6, + "total_downloads": 49181, + "min_param_count": 151148576, "sample_models": [ - "amazon/GKA-primed-HQwen3-8B-Instruct", - "amazon/Mamba2-primed-HQwen3-8B-Instruct", - "amazon/GDN-primed-HQwen3-8B-Instruct", - "amazon/GDN-primed-HQwen3-32B-Instruct", - "amazon/GKA-primed-HQwen3-32B-Instruct", - "amazon/BMOJOF-primed-HQwen3-8B-Instruct", - "amazon/GKA-primed-HQwen3-8B-Reasoner", - "amazon/GDN-primed-HQwen3-8B-Reasoner", - "amazon/GKA-primed-HQwen3-32B-Reasoner" - ] + "hugohrban/progen2-base", + "hugohrban/progen2-small", + "hugohrban/progen2-medium", + "hugohrban/progen2-large", + "hugohrban/progen2-small-mix7", + "hugohrban/progen2-oas" + ], + "relevancy_score": 47.6 }, { - "architecture_id": "BartForConditionalGeneration", - "total_models": 8, + "architecture_id": "HyenaDNAForCausalLM", + "total_models": 7, + "total_downloads": 36677, + "min_param_count": 450712, "sample_models": [ - "KomeijiForce/bart-large-emojilm", - "antalvdb/bart-base-spelling-nl", - "lmqg/bart-large-squad-qg", - "kengurukleo/deutsch_a2_transformer", - "shibing624/bart4csc-base-chinese", - "SkitCon/gec-spanish-BARTO-SYNTHETIC", - "Nargizi/screeve-lemmatizer", - "Tianlin668/MentalBART" - ] + "LongSafari/hyenadna-small-32k-seqlen-hf", + "LongSafari/hyenadna-medium-450k-seqlen-hf", + "LongSafari/hyenadna-large-1m-seqlen-hf", + "LongSafari/hyenadna-tiny-1k-seqlen-hf", + "LongSafari/hyenadna-medium-160k-seqlen-hf", + "LongSafari/hyenadna-tiny-16k-seqlen-d128-hf", + "LongSafari/hyenadna-tiny-1k-seqlen-d256-hf" + ], + "relevancy_score": 47.5 }, { - "architecture_id": "MambaForCausalLM", - "total_models": 8, + "architecture_id": "NemotronForCausalLM", + "total_models": 5, + "total_downloads": 59685, + "min_param_count": 2150720, "sample_models": [ - "state-spaces/mamba-130m-hf", - "state-spaces/mamba-2.8b-hf", - "state-spaces/mamba-1.4b-hf", - "state-spaces/mamba-370m-hf", - "state-spaces/mamba-790m-hf", - "NYTK/PULI-HuBA-mamba-130M", - "EchoLabs33/mamba-130m-hxq", - "TRI-ML/mamba-7b-rw" - ] + "nvidia/Nemotron-Mini-4B-Instruct", + "nvidia/Minitron-8B-Base", + "nvidia/Minitron-4B-Base", + "badaoui/tiny-random-NemotronForCausalLM", + "thhaus/nemotron3-8b" + ], + "relevancy_score": 47.4 }, { - "architecture_id": "Lfm2MoeForCausalLM", - "total_models": 8, + "architecture_id": "Glm4ForCausalLM", + "total_models": 7, + "total_downloads": 30337, + "min_param_count": 4854928, "sample_models": [ - "farbodtavakkoli/OTel-LLM-24B-IT", - "LiquidAI/LFM2-8B-A1B", - "LiquidAI/LFM2-24B-A2B", - "LiquidAI/LFM2-8B-A1B-ONNX", - "LiquidAI/LFM2-24B-A2B-ONNX", - "unsloth/LFM2-8B-A1B", - "huihui-ai/Huihui-LFM2-24B-A2B-abliterated", - "MuXodious/LFM2-8B-A1B-absolute-heresy-MPOA" - ] + "zai-org/GLM-4-9B-0414", + "zai-org/GLM-Z1-32B-0414", + "zai-org/GLM-Z1-9B-0414", + "zai-org/GLM-4-32B-0414", + "zai-org/GLM-4-32B-Base-0414", + "llmfan46/GLM-4-32B-0414-uncensored-heretic-v1", + "yujiepan/glm-4-tiny-random" + ], + "relevancy_score": 47.1 }, { - "architecture_id": "BloomModel", - "total_models": 8, + "architecture_id": "Eagle3DraftModel", + "total_models": 7, + "total_downloads": 24682, + "min_param_count": 522152832, "sample_models": [ - "bigscience/bigscience-small-testing", - "TurkuNLP/gpt3-finnish-small", - "TurkuNLP/gpt3-finnish-large", - "TurkuNLP/gpt3-finnish-13B", - "BelleGroup/BELLE-7B-2M", - "norallm/norbloom-7b-scratch", - "Muennighoff/bloom-tiny-random", - "TurkuNLP/gpt3-finnish-xl" - ] + "RedHatAI/gpt-oss-20b-speculator.eagle3", + "RedHatAI/gpt-oss-120b-speculator.eagle3", + "RedHatAI/Qwen3-30B-A3B-Thinking-2507-speculator.eagle3", + "RedHatAI/Qwen3-235B-A22B-Instruct-2507-speculator.eagle3", + "RedHatAI/Qwen3-30B-A3B-Instruct-2507-speculator.eagle3", + "RedHatAI/Qwen3-30B-A3B-speculator.eagle3", + "RedHatAI/Qwen3-32B-Thinking-speculator.eagle3" + ], + "relevancy_score": 46.6 }, { - "architecture_id": "ExaoneForCausalLM", + "architecture_id": "T5WithLMHeadModel", "total_models": 7, + "total_downloads": 24870, + "min_param_count": 222903936, "sample_models": [ - "LGAI-EXAONE/EXAONE-Deep-7.8B", - "LGAI-EXAONE/EXAONE-3.5-7.8B-Instruct", - "LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct", - "LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct", - "LGAI-EXAONE/EXAONE-3.5-32B-Instruct", - "LGAI-EXAONE/EXAONE-Deep-32B", - "LGAI-EXAONE/EXAONE-Deep-2.4B" - ] + "Salesforce/codet5-large", + "unicamp-dl/ptt5-base-portuguese-vocab", + "Salesforce/codet5-large-ntp-py", + "Rostlab/prot_t5_xl_bfd", + "unicamp-dl/ptt5-small-portuguese-vocab", + "gagan3012/k2t", + "unicamp-dl/ptt5-large-portuguese-vocab" + ], + "relevancy_score": 46.6 + }, + { + "architecture_id": "AquilaForCausalLM", + "total_models": 7, + "total_downloads": 19362, + "min_param_count": 6425376, + "sample_models": [ + "BAAI/AquilaChat2-7B", + "katuni4ka/tiny-random-aquilachat", + "katuni4ka/tiny-random-aquila2", + "BAAI/Aquila2-34B", + "BAAI/AquilaChat2-34B", + "BAAI/AquilaChat2-34B-16K", + "BAAI/Aquila2-70B-Expr" + ], + "relevancy_score": 46.1 }, { "architecture_id": "Zamba2ForCausalLM", "total_models": 7, + "total_downloads": 110981, + "min_param_count": 1215064704, "sample_models": [ "Zyphra/Zamba2-1.2B-instruct", "Zyphra/Zamba2-7B-Instruct", @@ -661,11 +917,14 @@ "Zyphra/Zamba2-2.7B-instruct", "EchoLabs33/zamba2-2.7b-instruct-hxq", "EchoLabs33/zamba2-7b-instruct-hxq" - ] + ], + "relevancy_score": 46.0 }, { "architecture_id": "LlamaForCausalLMEagle3", "total_models": 7, + "total_downloads": 18213, + "min_param_count": 208676608, "sample_models": [ "taobao-mnn/Qwen3-VL-8B-Instruct-Eagle3", "nvidia/gpt-oss-120b-Eagle3-long-context", @@ -674,109 +933,540 @@ "Zjcxy-SmartAI/Eagle3-Qwen3-8B-zh", "nvidia/gpt-oss-120b-Eagle3-throughput", "chankhavu/c2.eagle3-test" - ] + ], + "relevancy_score": 45.9 }, { - "architecture_id": "T5WithLMHeadModel", + "architecture_id": "XverseForCausalLM", "total_models": 7, + "total_downloads": 16554, + "min_param_count": 6459056, "sample_models": [ - "Salesforce/codet5-large", - "unicamp-dl/ptt5-base-portuguese-vocab", - "Salesforce/codet5-large-ntp-py", - "Rostlab/prot_t5_xl_bfd", - "unicamp-dl/ptt5-small-portuguese-vocab", - "gagan3012/k2t", - "unicamp-dl/ptt5-large-portuguese-vocab" - ] + "xverse/XVERSE-7B-Chat", + "katuni4ka/tiny-random-xverse", + "xverse/XVERSE-13B-256K", + "xverse/XVERSE-13B", + "xverse/XVERSE-65B-Chat", + "xverse/XVERSE-13B-Chat", + "xverse/XVERSE-7B" + ], + "relevancy_score": 45.7 }, { - "architecture_id": "Rwkv6ForCausalLM", - "total_models": 7, + "architecture_id": "ArceeForCausalLM", + "total_models": 4, + "total_downloads": 36570, + "min_param_count": 4129088, "sample_models": [ - "RWKV/v6-Finch-1B6-HF", - "RWKV/v6-Finch-7B-HF", - "RWKV/rwkv-6-world-1b6", - "RWKV/v6-Finch-14B-HF", - "RWKV/v6-Finch-3B-HF", - "RWKV/rwkv-6-world-7b", - "RWKV/rwkv-6-world-3b-v2.1" - ] + "arcee-ai/AFM-4.5B-Base", + "optimum-intel-internal-testing/tiny-random-ArceeForCausalLM", + "onnx-internal-testing/tiny-random-ArceeForCausalLM", + "arcee-ai/AFM-4.5B" + ], + "relevancy_score": 45.7 + }, + { + "architecture_id": "LlavaQwen2ForCausalLM", + "total_models": 5, + "total_downloads": 26319, + "min_param_count": 758833760, + "sample_models": [ + "qnguyen3/nanoLLaVA", + "apple/FastVLM-0.5B", + "apple/FastVLM-1.5B", + "apple/FastVLM-7B", + "FreedomIntelligence/HuatuoGPT-Vision-7B" + ], + "relevancy_score": 45.6 + }, + { + "architecture_id": "SDARForCausalLM", + "total_models": 6, + "total_downloads": 95550, + "min_param_count": 2031739904, + "sample_models": [ + "JetLM/SDAR-1.7B-Chat", + "JetLM/SDAR-8B-Chat-b32", + "JetLM/SDAR-8B-Chat", + "JetLM/SDAR-1.7B-Chat-b32", + "JetLM/SDAR-4B-Chat", + "JetLM/SDAR-4B-Chat-b32" + ], + "relevancy_score": 45.0 + }, + { + "architecture_id": "BitNetForCausalLM", + "total_models": 3, + "total_downloads": 25978, + "min_param_count": 849787090, + "sample_models": [ + "microsoft/bitnet-b1.58-2B-4T", + "microsoft/bitnet-b1.58-2B-4T-bf16", + "iSolver-AI/FEnet" + ], + "relevancy_score": 44.4 + }, + { + "architecture_id": "PldrllmForCausalLM", + "total_models": 5, + "total_downloads": 13606, + "min_param_count": 109689362, + "sample_models": [ + "fromthesky/PLDR-LLM-v51-SOC-110M-5", + "fromthesky/PLDR-LLM-v51-SOC-110M-2", + "fromthesky/PLDR-LLM-v51-SOC-110M-4", + "fromthesky/PLDR-LLM-v51-SOC-110M-3", + "fromthesky/PLDR-LLM-v51-SOC-110M-1" + ], + "relevancy_score": 44.1 + }, + { + "architecture_id": "DeciLMForCausalLM", + "total_models": 14, + "total_downloads": 262843, + "min_param_count": 7043551232, + "sample_models": [ + "nvidia/Llama-3_3-Nemotron-Super-49B-v1_5", + "nvidia/Llama-3_3-Nemotron-Super-49B-v1", + "nvidia/Llama-3_3-Nemotron-Super-49B-v1_5-NVFP4", + "nvidia/Llama-3_1-Nemotron-Ultra-253B-v1", + "ConicCat/Llama3_3-Nemo-Super-Writer-49B", + "nvidia/Llama-3_1-Nemotron-51B-Instruct", + "FriendliAI/Llama-3_3-Nemotron-Super-49B-v1_5", + "FriendliAI/Llama-3_1-Nemotron-Ultra-253B-v1", + "NewstaR/Porpoise-6b-instruct", + "nvidia/Llama-3_1-Nemotron-Ultra-253B-CPT-v1" + ], + "relevancy_score": 44.0 + }, + { + "architecture_id": "MoAMetricLM", + "total_models": 5, + "total_downloads": 12058, + "min_param_count": 69130608, + "sample_models": [ + "reaperdoesntknow/MoA-150M", + "reaperdoesntknow/MoA-400M", + "reaperdoesntknow/MoA-155M", + "reaperdoesntknow/MoA-100M", + "reaperdoesntknow/DiscoverLM-70M" + ], + "relevancy_score": 43.8 + }, + { + "architecture_id": "Llama4ForCausalLM", + "total_models": 3, + "total_downloads": 20475, + "min_param_count": 3269144, + "sample_models": [ + "trl-internal-testing/tiny-Llama4ForCausalLM", + "pruna-test/test-save-tiny-random-llama4-smashed", + "facebook/MobileLLM-R1.5-360M" + ], + "relevancy_score": 43.8 + }, + { + "architecture_id": "MiniMaxM2ForCausalLM", + "total_models": 15, + "total_downloads": 998765, + "min_param_count": 18581099008, + "sample_models": [ + "MiniMaxAI/MiniMax-M2.5", + "cerebras/MiniMax-M2.1-REAP-139B-A10B", + "MiniMaxAI/MiniMax-M2", + "MiniMaxAI/MiniMax-M2.1", + "nvidia/MiniMax-M2.5-NVFP4", + "cerebras/MiniMax-M2.5-REAP-139B-A10B", + "amd/MiniMax-M2.5-MXFP4", + "saricles/MiniMax-M2.5-REAP-172B-A10B-NVFP4-GB10", + "aspctu/MiniMax-M2.5", + "amd/MiniMax-M2.1-MXFP4" + ], + "relevancy_score": 43.6 + }, + { + "architecture_id": "MBartForConditionalGeneration", + "total_models": 6, + "total_downloads": 7278, + "min_param_count": 379691717, + "sample_models": [ + "Pravopysnyk/best-unlp", + "DeepPavlov/mbart-large-50-ru-persona-chat", + "sn4kebyt3/ru-bart-large", + "MRNH/mbart-italian-grammar-corrector", + "MRNH/mbart-german-grammar-corrector", + "MRNH/mbart-russian-grammar-corrector" + ], + "relevancy_score": 43.3 + }, + { + "architecture_id": "DogeForCausalLM", + "total_models": 6, + "total_downloads": 7189, + "min_param_count": 13118728, + "sample_models": [ + "SmallDoge/Doge-320M", + "SmallDoge/Doge-20M", + "SmallDoge/Doge-160M", + "SmallDoge/Doge-60M", + "SmallDoge/Doge-120M-MoE", + "SmallDoge/Doge-20M-MoE" + ], + "relevancy_score": 43.3 + }, + { + "architecture_id": "NemotronFlashForCausalLM", + "total_models": 2, + "total_downloads": 21363, + "min_param_count": 965389440, + "sample_models": [ + "nvidia/Nemotron-Flash-3B", + "nvidia/Nemotron-Flash-1B" + ], + "relevancy_score": 43.3 + }, + { + "architecture_id": "Exaone4ForCausalLM", + "total_models": 3, + "total_downloads": 86500, + "min_param_count": 1279391488, + "sample_models": [ + "LGAI-EXAONE/EXAONE-4.0-1.2B", + "LGAI-EXAONE/EXAONE-4.0.1-32B", + "LGAI-EXAONE/EXAONE-4.0-32B" + ], + "relevancy_score": 43.0 + }, + { + "architecture_id": "MiniMaxForCausalLM", + "total_models": 2, + "total_downloads": 16757, + "min_param_count": 231006264, + "sample_models": [ + "MiniMaxAI/MiniMax-Text-01-hf", + "hyper-accel/tiny-random-minimax" + ], + "relevancy_score": 42.8 + }, + { + "architecture_id": "EchoForCausalLM", + "total_models": 2, + "total_downloads": 14462, + "min_param_count": 114687488, + "sample_models": [ + "ethicalabs/Echo-DSRN-486M-v0.7.6-SFT", + "ethicalabs/Echo-DSRN-114M-Base" + ], + "relevancy_score": 42.5 + }, + { + "architecture_id": "Mistral3ForConditionalGeneration", + "total_models": 6, + "total_downloads": 175114, + "min_param_count": 4251743232, + "sample_models": [ + "farbodtavakkoli/OTel-LLM-3B-IT", + "RedHatAI/Mistral-Small-3.2-24B-Instruct-2506-NVFP4", + "ArmGPT/ArmenianGPT-1.0-3B", + "JANGQ-AI/Mistral-Small-4-119B-A6B-JANG_2L", + "odytrice/kenichi-flash", + "JANGQ-AI/Mistral-Small-4-119B-A6B-JANG_6M" + ], + "relevancy_score": 42.4 + }, + { + "architecture_id": "IlamaForCausalLM", + "total_models": 1, + "total_downloads": 106257, + "min_param_count": 1235814400, + "sample_models": [ + "hmellor/Ilama-3.2-1B" + ], + "relevancy_score": 42.3 + }, + { + "architecture_id": "AraGPT2LMHeadModel", + "total_models": 3, + "total_downloads": 9601, + "min_param_count": 829369856, + "sample_models": [ + "QCRI/Fanar-2-Diwan", + "aubmindlab/aragpt2-mega", + "aubmindlab/aragpt2-large" + ], + "relevancy_score": 42.2 + }, + { + "architecture_id": "StarVectorForCausalLM", + "total_models": 2, + "total_downloads": 74247, + "min_param_count": 1434095620, + "sample_models": [ + "starvector/starvector-1b-im2svg", + "starvector/starvector-8b-im2svg" + ], + "relevancy_score": 42.1 + }, + { + "architecture_id": "DbrxForCausalLM", + "total_models": 2, + "total_downloads": 12324, + "min_param_count": 1612456, + "sample_models": [ + "trl-internal-testing/tiny-DbrxForCausalLM", + "katuni4ka/tiny-random-dbrx" + ], + "relevancy_score": 42.1 + }, + { + "architecture_id": "RITAModelForCausalLM", + "total_models": 3, + "total_downloads": 9058, + "min_param_count": 85096320, + "sample_models": [ + "lightonai/RITA_s", + "lightonai/RITA_xl", + "lightonai/RITA_l" + ], + "relevancy_score": 42.0 + }, + { + "architecture_id": "ModernBertForSequenceClassification", + "total_models": 1, + "total_downloads": 15438, + "min_param_count": 149609478, + "sample_models": [ + "opendatalab/meta-rater-professionalism-rating" + ], + "relevancy_score": 42.0 + }, + { + "architecture_id": "LlavaLlamaForCausalLM", + "total_models": 18, + "total_downloads": 33414, + "min_param_count": 7466764288, + "sample_models": [ + "LanguageBind/Video-LLaVA-7B", + "wisdomik/Quilt-Llava-v1.5-7b", + "liuhaotian/llava-llama-2-13b-chat-lightning-preview", + "lmms-lab/llama3-llava-next-8b", + "mmaaz60/LLaVA-7B-Lightening-v1-1", + "microsoft/llava-med-7b-delta", + "deepcs233/VisCoT-7b-336", + "ManishThota/Ollama_Video_llama_7B", + "EricPolaris/Quilt-Llava-v1.5-7b", + "liuhaotian/LLaVA-Lightning-7B-delta-v1-1" + ], + "relevancy_score": 41.8 + }, + { + "architecture_id": "Llama4ForConditionalGeneration", + "total_models": 3, + "total_downloads": 8252, + "min_param_count": 6686880, + "sample_models": [ + "yujiepan/llama-4-tiny-random", + "RedHatAI/Llama-4-Scout-17B-16E-Instruct-NVFP4", + "RedHatAI/Llama-4-Maverick-17B-128E-Instruct-NVFP4" + ], + "relevancy_score": 41.8 + }, + { + "architecture_id": "GlmForCausalLM", + "total_models": 4, + "total_downloads": 35472, + "min_param_count": 1593427968, + "sample_models": [ + "zai-org/glm-4-9b-chat-hf", + "zai-org/glm-4-9b-hf", + "zai-org/glm-edge-4b-chat", + "zai-org/glm-edge-1.5b-chat" + ], + "relevancy_score": 41.7 + }, + { + "architecture_id": "Plamo2ForCausalLM", + "total_models": 1, + "total_downloads": 81533, + "min_param_count": 1291441920, + "sample_models": [ + "pfnet/plamo-2-1b" + ], + "relevancy_score": 41.7 + }, + { + "architecture_id": "LLaMAForCausalLM", + "total_models": 12, + "total_downloads": 21882, + "min_param_count": 6738425856, + "sample_models": [ + "maicomputer/alpaca-13b", + "Enoch/llama-65b-hf", + "mncai/chatdoctor", + "AdaptLLM/law-LLM", + "Nitish-Garikoti/finance-LLM", + "boboto/LLaMA-65B-HF", + "AdaptLLM/finance-LLM", + "AdaptLLM/medicine-LLM", + "Rardilit/Panther_v1", + "James-WYang/BigTranslate" + ], + "relevancy_score": 41.3 + }, + { + "architecture_id": "OLMoForCausalLM", + "total_models": 6, + "total_downloads": 16631, + "min_param_count": 1176764416, + "sample_models": [ + "allenai/OLMo-7B-Instruct", + "allenai/OLMo-7B", + "allenai/OLMo-1B", + "allenai/OLMo-7B-0424", + "allenai/OLMo-7B-Twin-2T", + "allenai/OLMo-7B-SFT" + ], + "relevancy_score": 41.2 + }, + { + "architecture_id": "MolformerForCausalLM", + "total_models": 2, + "total_downloads": 7447, + "min_param_count": 46805760, + "sample_models": [ + "ibm-research/GP-MoLFormer-Uniq", + "ralyn/NPComposer-v2" + ], + "relevancy_score": 41.0 + }, + { + "architecture_id": "Starcoder2ForCausalLM", + "total_models": 5, + "total_downloads": 117613, + "min_param_count": 3030371328, + "sample_models": [ + "bigcode/starcoder2-3b", + "bigcode/starcoder2-7b", + "bigcode/starcoder2-15b", + "bigcode/starcoder2-15b-instruct-v0.1", + "dphn/dolphincoder-starcoder2-15b" + ], + "relevancy_score": 40.9 + }, + { + "architecture_id": "GLAForCausalLM", + "total_models": 3, + "total_downloads": 5517, + "min_param_count": 341707776, + "sample_models": [ + "fla-hub/gla-340M-15B", + "fla-hub/gla-1.3B-100B", + "fla-hub/gla-2.7B-100B" + ], + "relevancy_score": 40.9 + }, + { + "architecture_id": "MptForCausalLM", + "total_models": 3, + "total_downloads": 4608, + "min_param_count": 405032, + "sample_models": [ + "yujiepan/mpt-tiny-random", + "explosion-testing/mpt-test", + "team-lucid/mptk-1b" + ], + "relevancy_score": 40.5 + }, + { + "architecture_id": "Mamba2ForCausalLM", + "total_models": 2, + "total_downloads": 6072, + "min_param_count": 316164736, + "sample_models": [ + "deqing/mamba2-300M-v5-mamba2", + "EchoLabs33/mamba2-1.3b-hxq" + ], + "relevancy_score": 40.5 }, { - "architecture_id": "GlmMoeDsaForCausalLM", - "total_models": 6, + "architecture_id": "NandiForCausalLM", + "total_models": 1, + "total_downloads": 7416, + "min_param_count": 153412928, "sample_models": [ - "zai-org/GLM-5", - "nvidia/GLM-5-NVFP4", - "zai-org/GLM-5.1", - "cs2764/GLM-5-abliterated-dq4-mlx", - "0xSero/GLM-5-REAP-381B", - "cs2764/GLM-5-abliterated-dq3-mlx" - ] + "Rta-AILabs/Nandi-Mini-150M" + ], + "relevancy_score": 40.4 }, { - "architecture_id": "Qwen2MoeForCausalLM", - "total_models": 6, + "architecture_id": "LLaDAModelLM", + "total_models": 4, + "total_downloads": 697086, + "min_param_count": 8015581184, "sample_models": [ - "Qwen/Qwen1.5-MoE-A2.7B", - "Qwen/Qwen1.5-MoE-A2.7B-Chat", - "Qwen/Qwen2-57B-A14B-Instruct", - "Qwen/Qwen2-57B-A14B", - "katuni4ka/tiny-random-qwen1.5-moe", - "yujiepan/qwen1.5-moe-tiny-random" - ] + "GSAI-ML/LLaDA-8B-Instruct", + "GSAI-ML/LLaDA-8B-Base", + "GSAI-ML/LLaDA-1.5", + "d3LLM/d3LLM_LLaDA" + ], + "relevancy_score": 40.3 }, { - "architecture_id": "DreamModel", - "total_models": 6, + "architecture_id": "DuchifatCore", + "total_models": 3, + "total_downloads": 3952, + "min_param_count": 136763904, "sample_models": [ - "Dream-org/Dream-v0-Instruct-7B", - "Dream-org/Dream-v0-Base-7B", - "Dream-org/Dream-Coder-v0-Instruct-7B", - "Zigeng/dParallel_Dream_7B_Instruct", - "Dream-org/Dream-Coder-v0-Base-7B", - "Dream-org/DreamOn-v0-7B" - ] + "Raziel1234/Duchifat-2", + "razielAI/Duchifat-2.1-Instruct", + "TopAI-1/Duchifat-2-Instruct" + ], + "relevancy_score": 40.2 }, { - "architecture_id": "Phi3VForCausalLM", - "total_models": 6, + "architecture_id": "BertLMHeadModel", + "total_models": 2, + "total_downloads": 4554, + "min_param_count": 184474880, "sample_models": [ - "microsoft/Phi-3-vision-128k-instruct", - "TIGER-Lab/VLM2Vec-Full", - "yujiepan/phi-3-vision-tiny-random", - "furonghuang-lab/tracevla_phi3v", - "Desm0nt/Phi-3-HornyVision-128k-instruct", - "failspy/Phi-3-vision-128k-instruct-abliterated-alpha" - ] + "dicta-il/BEREL_3.0", + "hf-tiny-model-private/tiny-random-BertLMHeadModel" + ], + "relevancy_score": 39.9 }, { - "architecture_id": "SDARForCausalLM", - "total_models": 6, + "architecture_id": "Lfm2MoeForCausalLM", + "total_models": 8, + "total_downloads": 173491, + "min_param_count": 8339929856, "sample_models": [ - "JetLM/SDAR-1.7B-Chat", - "JetLM/SDAR-8B-Chat-b32", - "JetLM/SDAR-8B-Chat", - "JetLM/SDAR-1.7B-Chat-b32", - "JetLM/SDAR-4B-Chat", - "JetLM/SDAR-4B-Chat-b32" - ] + "farbodtavakkoli/OTel-LLM-24B-IT", + "LiquidAI/LFM2-8B-A1B", + "LiquidAI/LFM2-24B-A2B", + "LiquidAI/LFM2-8B-A1B-ONNX", + "LiquidAI/LFM2-24B-A2B-ONNX", + "unsloth/LFM2-8B-A1B", + "huihui-ai/Huihui-LFM2-24B-A2B-abliterated", + "MuXodious/LFM2-8B-A1B-absolute-heresy-MPOA" + ], + "relevancy_score": 39.5 }, { - "architecture_id": "HyenaDNAForCausalLM", - "total_models": 6, + "architecture_id": "BitnetForCausalLM", + "total_models": 2, + "total_downloads": 3581, + "min_param_count": 728843904, "sample_models": [ - "LongSafari/hyenadna-small-32k-seqlen-hf", - "LongSafari/hyenadna-medium-450k-seqlen-hf", - "LongSafari/hyenadna-large-1m-seqlen-hf", - "LongSafari/hyenadna-tiny-1k-seqlen-hf", - "LongSafari/hyenadna-medium-160k-seqlen-hf", - "LongSafari/hyenadna-tiny-16k-seqlen-d128-hf" - ] + "1bitLLM/bitnet_b1_58-large", + "1bitLLM/bitnet_b1_58-3B" + ], + "relevancy_score": 39.4 }, { "architecture_id": "AfmoeForCausalLM", "total_models": 6, + "total_downloads": 44217, + "min_param_count": 6120003328, "sample_models": [ "arcee-ai/Trinity-Nano-Preview", "arcee-ai/Trinity-Large-Thinking", @@ -784,2486 +1474,3277 @@ "arcee-ai/Trinity-Nano-Base", "arcee-ai/Trinity-Mini-Base", "arcee-ai/Trinity-Large-Preview" - ] + ], + "relevancy_score": 39.3 }, { - "architecture_id": "AquilaForCausalLM", - "total_models": 6, + "architecture_id": "GatedDeltaNetForCausalLM", + "total_models": 1, + "total_downloads": 4063, + "min_param_count": 317524480, "sample_models": [ - "BAAI/AquilaChat2-7B", - "katuni4ka/tiny-random-aquilachat", - "katuni4ka/tiny-random-aquila2", - "BAAI/Aquila2-34B", - "BAAI/AquilaChat2-34B", - "BAAI/AquilaChat2-34B-16K" - ] + "deqing/gdn-300M-v5-gdn" + ], + "relevancy_score": 39.1 }, { - "architecture_id": "OLMoForCausalLM", - "total_models": 6, + "architecture_id": "RecurrentGemmaForCausalLM", + "total_models": 3, + "total_downloads": 13140, + "min_param_count": 2682862080, "sample_models": [ - "allenai/OLMo-7B-Instruct", - "allenai/OLMo-7B", - "allenai/OLMo-1B", - "allenai/OLMo-7B-0424", - "allenai/OLMo-7B-Twin-2T", - "allenai/OLMo-7B-SFT" - ] + "google/recurrentgemma-2b", + "google/recurrentgemma-2b-it", + "google/recurrentgemma-9b" + ], + "relevancy_score": 38.9 }, { - "architecture_id": "DogeForCausalLM", - "total_models": 6, + "architecture_id": "RecursiveLanguageModel", + "total_models": 1, + "total_downloads": 3565, + "min_param_count": 198464806, "sample_models": [ - "SmallDoge/Doge-320M", - "SmallDoge/Doge-20M", - "SmallDoge/Doge-160M", - "SmallDoge/Doge-60M", - "SmallDoge/Doge-120M-MoE", - "SmallDoge/Doge-20M-MoE" - ] + "Girinath11/recursive-language-model-198m" + ], + "relevancy_score": 38.8 }, { - "architecture_id": "SmolLM3ForCausalLM", - "total_models": 5, + "architecture_id": "RWForCausalLM", + "total_models": 10, + "total_downloads": 11287, + "min_param_count": 6854619456, "sample_models": [ - "HuggingFaceTB/SmolLM3-3B", - "HuggingFaceTB/SmolLM3-3B-Base", - "optimum-internal-testing/tiny-random-SmolLM3ForCausalLM", - "unsloth/SmolLM3-3B", - "onnx-internal-testing/tiny-random-SmolLM3ForCausalLM" - ] + "projecte-aina/aguila-7b", + "lightonai/alfred-40b-1023", + "explosion-testing/refined-web-model-test", + "vilm/vulture-40b", + "h2oai/h2ogpt-gm-oasst1-en-2048-falcon-7b-v2", + "nomic-ai/gpt4all-falcon", + "h2oai/h2ogpt-gm-oasst1-en-2048-falcon-7b-v3", + "OpenAssistant/falcon-40b-sft-top1-560", + "QuixiAI/WizardLM-Uncensored-Falcon-40b", + "mrm8488/falcoder-7b" + ], + "relevancy_score": 38.7 }, { - "architecture_id": "XLNetLMHeadModel", - "total_models": 5, + "architecture_id": "T5EncoderModel", + "total_models": 1, + "total_downloads": 124383, + "min_param_count": 4762310656, "sample_models": [ - "xlnet/xlnet-base-cased", - "xlnet/xlnet-large-cased", - "hfl/chinese-xlnet-base", - "sshleifer/tiny-xlnet-base-cased", - "textattack/xlnet-base-cased-imdb" - ] + "XLabs-AI/xflux_text_encoders" + ], + "relevancy_score": 38.7 }, { - "architecture_id": "BioGptForCausalLM", - "total_models": 5, + "architecture_id": "AprielForCausalLM", + "total_models": 1, + "total_downloads": 113509, + "min_param_count": 4832071680, "sample_models": [ - "microsoft/biogpt", - "microsoft/BioGPT-Large", - "microsoft/BioGPT-Large-PubMedQA", - "hf-tiny-model-private/tiny-random-BioGptForCausalLM", - "zequnl/molxpt" - ] + "ServiceNow-AI/Apriel-5B-Instruct" + ], + "relevancy_score": 38.5 }, { - "architecture_id": "Mistral3ForConditionalGeneration", - "total_models": 5, + "architecture_id": "LLM", + "total_models": 1, + "total_downloads": 3032, + "min_param_count": 497145984, "sample_models": [ - "farbodtavakkoli/OTel-LLM-3B-IT", - "RedHatAI/Mistral-Small-3.2-24B-Instruct-2506-NVFP4", - "ArmGPT/ArmenianGPT-1.0-3B", - "JANGQ-AI/Mistral-Small-4-119B-A6B-JANG_2L", - "odytrice/kenichi-flash" - ] + "rudyon/linnet-497M" + ], + "relevancy_score": 38.4 }, { - "architecture_id": "LLaDA2MoeModelLM", - "total_models": 5, + "architecture_id": "SwarmForCausalLM", + "total_models": 1, + "total_downloads": 2951, + "min_param_count": 52729731, "sample_models": [ - "inclusionAI/LLaDA2.1-flash", - "inclusionAI/LLaDA2.0-mini", - "inclusionAI/LLaDA2.1-mini", - "inclusionAI/LLaDA2.0-mini-CAP", - "inclusionAI/LLaDA2.0-flash" - ] + "reaperdoesntknow/SAGI" + ], + "relevancy_score": 38.4 }, { - "architecture_id": "FalconMambaForCausalLM", - "total_models": 5, + "architecture_id": "SpatialLMQwenForCausalLM", + "total_models": 1, + "total_downloads": 2727, + "min_param_count": 603511168, "sample_models": [ - "trl-internal-testing/tiny-FalconMambaForCausalLM", - "tiiuae/falcon-mamba-7b-instruct", - "tiiuae/falcon-mamba-7b", - "tiiuae/falcon-mamba-tiny-dev", - "tiiuae/Falcon3-Mamba-7B-Instruct" - ] + "manycore-research/SpatialLM1.1-Qwen-0.5B" + ], + "relevancy_score": 38.2 }, { - "architecture_id": "Eagle3Speculator", - "total_models": 5, + "architecture_id": "SliderGPT", + "total_models": 2, + "total_downloads": 1950, + "min_param_count": 47420936, "sample_models": [ - "RedHatAI/Qwen3-8B-speculator.eagle3", - "RedHatAI/Llama-3.1-8B-Instruct-speculator.eagle3", - "RedHatAI/Llama-3.3-70B-Instruct-speculator.eagle3", - "RedHatAI/Qwen3-32B-speculator.eagle3", - "RedHatAI/Qwen3-14B-speculator.eagle3" - ] + "c-bone/CrystaLLM-pi_Mattergen-XRD", + "c-bone/CrystaLLM-pi_COD-XRD" + ], + "relevancy_score": 38.0 }, { - "architecture_id": "NemotronForCausalLM", - "total_models": 5, + "architecture_id": "MiniMindForCausalLM", + "total_models": 2, + "total_downloads": 1929, + "min_param_count": 38840960, "sample_models": [ - "nvidia/Nemotron-Mini-4B-Instruct", - "nvidia/Minitron-8B-Base", - "nvidia/Minitron-4B-Base", - "badaoui/tiny-random-NemotronForCausalLM", - "thhaus/nemotron3-8b" - ] + "yiwenX/MiniMind-MoE-640-120M", + "chujiamo/baiheng_0405" + ], + "relevancy_score": 38.0 }, { - "architecture_id": "ProGenForCausalLM", - "total_models": 5, + "architecture_id": "AV2TextForConditionalGeneration", + "total_models": 1, + "total_downloads": 2542, + "min_param_count": 480465000, "sample_models": [ - "hugohrban/progen2-base", - "hugohrban/progen2-small", - "hugohrban/progen2-medium", - "hugohrban/progen2-large", - "hugohrban/progen2-small-mix7" - ] + "nguyenvulebinh/AV-HuBERT-MuAViC-en" + ], + "relevancy_score": 38.0 }, { - "architecture_id": "Glm4ForCausalLM", - "total_models": 5, + "architecture_id": "GPTJXMoEForCausalLM", + "total_models": 1, + "total_downloads": 2496, + "min_param_count": 489915648, "sample_models": [ - "zai-org/GLM-4-9B-0414", - "zai-org/GLM-Z1-32B-0414", - "zai-org/GLM-Z1-9B-0414", - "zai-org/GLM-4-32B-0414", - "zai-org/GLM-4-32B-Base-0414" - ] + "Aletheia-ng/SabiYarn_MoE_translate" + ], + "relevancy_score": 38.0 }, { - "architecture_id": "Eagle3DraftModel", - "total_models": 5, + "architecture_id": "DreamModel", + "total_models": 6, + "total_downloads": 142972, + "min_param_count": 7615616512, "sample_models": [ - "RedHatAI/gpt-oss-20b-speculator.eagle3", - "RedHatAI/gpt-oss-120b-speculator.eagle3", - "RedHatAI/Qwen3-30B-A3B-Thinking-2507-speculator.eagle3", - "RedHatAI/Qwen3-235B-A22B-Instruct-2507-speculator.eagle3", - "RedHatAI/Qwen3-30B-A3B-Instruct-2507-speculator.eagle3" - ] + "Dream-org/Dream-v0-Instruct-7B", + "Dream-org/Dream-v0-Base-7B", + "Dream-org/Dream-Coder-v0-Instruct-7B", + "Zigeng/dParallel_Dream_7B_Instruct", + "Dream-org/Dream-Coder-v0-Base-7B", + "Dream-org/DreamOn-v0-7B" + ], + "relevancy_score": 37.9 }, { - "architecture_id": "LlavaQwen2ForCausalLM", - "total_models": 5, + "architecture_id": "PenguinVLQwen3ForCausalLM", + "total_models": 2, + "total_downloads": 10860, + "min_param_count": 2167941120, "sample_models": [ - "qnguyen3/nanoLLaVA", - "apple/FastVLM-0.5B", - "apple/FastVLM-1.5B", - "apple/FastVLM-7B", - "FreedomIntelligence/HuatuoGPT-Vision-7B" - ] + "tencent/Penguin-VL-8B", + "tencent/Penguin-VL-2B" + ], + "relevancy_score": 37.8 }, { - "architecture_id": "JAISLMHeadModel", - "total_models": 5, + "architecture_id": "BD3LM", + "total_models": 2, + "total_downloads": 1804, + "min_param_count": 169627250, "sample_models": [ - "inceptionai/jais-13b-chat", - "katuni4ka/tiny-random-jais", - "inceptionai/jais-family-30b-8k", - "inceptionai/jais-13b", - "inceptionai/jais-family-13b-chat" - ] + "kuleshov-group/bd3lm-owt-block_size4", + "kuleshov-group/bd3lm-owt-block_size16" + ], + "relevancy_score": 37.8 }, { - "architecture_id": "MoAMetricLM", - "total_models": 5, + "architecture_id": "BlenderbotForConditionalGeneration", + "total_models": 1, + "total_downloads": 2199, + "min_param_count": 364810568, + "sample_models": [ + "thu-coai/blenderbot-400M-esconv" + ], + "relevancy_score": 37.7 + }, + { + "architecture_id": "Autoencoder", + "total_models": 1, + "total_downloads": 2171, + "min_param_count": 75832064, "sample_models": [ - "reaperdoesntknow/MoA-150M", - "reaperdoesntknow/MoA-400M", - "reaperdoesntknow/MoA-155M", - "reaperdoesntknow/MoA-100M", - "reaperdoesntknow/DiscoverLM-70M" - ] + "cccczshao/CALM-Autoencoder" + ], + "relevancy_score": 37.7 }, { - "architecture_id": "PldrllmForCausalLM", - "total_models": 5, + "architecture_id": "HCXVisionV2ForCausalLM", + "total_models": 2, + "total_downloads": 357762, + "min_param_count": 10741664520, "sample_models": [ - "fromthesky/PLDR-LLM-v51-SOC-110M-5", - "fromthesky/PLDR-LLM-v51-SOC-110M-2", - "fromthesky/PLDR-LLM-v51-SOC-110M-4", - "fromthesky/PLDR-LLM-v51-SOC-110M-3", - "fromthesky/PLDR-LLM-v51-SOC-110M-1" - ] + "naver-hyperclovax/HyperCLOVAX-SEED-Omni-8B", + "naver-hyperclovax/HyperCLOVAX-SEED-Think-32B" + ], + "relevancy_score": 37.6 }, { - "architecture_id": "MBartForConditionalGeneration", - "total_models": 5, + "architecture_id": "FusionInDecoderForConditionalGeneration", + "total_models": 1, + "total_downloads": 2090, + "min_param_count": 247577856, "sample_models": [ - "Pravopysnyk/best-unlp", - "DeepPavlov/mbart-large-50-ru-persona-chat", - "sn4kebyt3/ru-bart-large", - "MRNH/mbart-italian-grammar-corrector", - "MRNH/mbart-german-grammar-corrector" - ] + "Intel/fid_flan_t5_base_nq" + ], + "relevancy_score": 37.6 }, { - "architecture_id": "PhiMoEForCausalLM", - "total_models": 4, + "architecture_id": "EveMoEForCausalLM", + "total_models": 1, + "total_downloads": 2088, + "min_param_count": 271970816, "sample_models": [ - "microsoft/Phi-tiny-MoE-instruct", - "microsoft/Phi-mini-MoE-instruct", - "microsoft/Phi-3.5-MoE-instruct", - "optimum-intel-internal-testing/phi-3.5-moe-tiny-random" - ] + "anthonym21/Eve-2-MoE-IT-272M" + ], + "relevancy_score": 37.6 }, { - "architecture_id": "LlavaQwenForCausalLM", - "total_models": 4, + "architecture_id": "Plamo3ForCausalLM", + "total_models": 1, + "total_downloads": 12440, + "min_param_count": 2603344384, "sample_models": [ - "lmms-lab/llava-onevision-qwen2-7b-ov", - "lmms-lab/llava-onevision-qwen2-0.5b-ov", - "lmms-lab/llava-onevision-qwen2-0.5b-si", - "lmms-lab/llava-onevision-qwen2-7b-si" - ] + "pfnet/plamo-3-nict-2b-base" + ], + "relevancy_score": 37.5 }, { - "architecture_id": "Starcoder2ForCausalLM", - "total_models": 4, + "architecture_id": "LIMEForCausalLM", + "total_models": 1, + "total_downloads": 2001, + "min_param_count": 984405504, "sample_models": [ - "bigcode/starcoder2-3b", - "bigcode/starcoder2-7b", - "bigcode/starcoder2-15b", - "bigcode/starcoder2-15b-instruct-v0.1" - ] + "anarlavrenov/lime-1b-instruct" + ], + "relevancy_score": 37.5 }, { - "architecture_id": "GlmForCausalLM", - "total_models": 4, + "architecture_id": "ModernBertForMaskedLM", + "total_models": 1, + "total_downloads": 1979, + "min_param_count": 590367063, "sample_models": [ - "zai-org/glm-4-9b-chat-hf", - "zai-org/glm-4-9b-hf", - "zai-org/glm-edge-4b-chat", - "zai-org/glm-edge-1.5b-chat" - ] + "JorgeVanco/diffusionGPT" + ], + "relevancy_score": 37.5 }, { - "architecture_id": "OuroForCausalLM", - "total_models": 4, + "architecture_id": "MoEGPTForCausalLM", + "total_models": 1, + "total_downloads": 1907, + "min_param_count": 149603328, "sample_models": [ - "ByteDance/Ouro-1.4B", - "ByteDance/Ouro-2.6B-Thinking", - "ByteDance/Ouro-2.6B", - "ByteDance/Ouro-1.4B-Thinking" - ] + "arnomatic/german-moe-gpt-v8-pretrained" + ], + "relevancy_score": 37.4 }, { - "architecture_id": "SeedOssForCausalLM", - "total_models": 4, + "architecture_id": "TransformerForCausalLM", + "total_models": 1, + "total_downloads": 10895, + "min_param_count": 1364297728, "sample_models": [ - "ByteDance-Seed/Seed-OSS-36B-Instruct", - "NousResearch/Hermes-4.3-36B", - "ByteDance-Seed/Seed-OSS-36B-Base", - "mratsim/Seed-OSS-36B-Instruct-NVFP4" - ] + "fla-hub/transformer-1.3B-100B" + ], + "relevancy_score": 37.3 }, { - "architecture_id": "ArceeForCausalLM", - "total_models": 4, + "architecture_id": "LSTMForCausalLM", + "total_models": 1, + "total_downloads": 1767, + "min_param_count": 164921344, "sample_models": [ - "arcee-ai/AFM-4.5B-Base", - "optimum-intel-internal-testing/tiny-random-ArceeForCausalLM", - "onnx-internal-testing/tiny-random-ArceeForCausalLM", - "arcee-ai/AFM-4.5B" - ] + "deqing/lstm-window-4-v5" + ], + "relevancy_score": 37.2 }, { - "architecture_id": "BailingMoeV2ForCausalLM", - "total_models": 4, + "architecture_id": "NanoChatForCausalLM", + "total_models": 3, + "total_downloads": 5694, + "min_param_count": 2217082880, "sample_models": [ - "inclusionAI/Ling-mini-2.0", - "inclusionAI/Ling-1T", - "inclusionAI/Ring-mini-2.0", - "inclusionAI/Ling-flash-2.0" - ] + "Twobombs/nanochat-d34-sft-hf", + "pankajmathur/nanochat-d34-sft-hf", + "Nekochu/nanochat-d24" + ], + "relevancy_score": 37.0 }, { - "architecture_id": "MobilintLlamaForCausalLM", - "total_models": 4, + "architecture_id": "Moondream", + "total_models": 1, + "total_downloads": 9789, + "min_param_count": 1857482608, "sample_models": [ - "mobilint/Llama-3.2-3B-Instruct", - "mobilint/Llama-3.2-1B-Instruct", - "mobilint/Llama-3.1-8B-Instruct", - "mobilint/HyperCLOVAX-SEED-Text-Instruct-1.5B" - ] + "vikhyatk/moondream1" + ], + "relevancy_score": 37.0 }, { - "architecture_id": "MobilintQwen2ForCausalLM", - "total_models": 4, + "architecture_id": "Qwen3ASRForConditionalGeneration", + "total_models": 2, + "total_downloads": 1198, + "min_param_count": 782426112, "sample_models": [ - "mobilint/Qwen2.5-1.5B-Instruct", - "mobilint/Qwen2.5-0.5B-Instruct", - "mobilint/Qwen2.5-3B-Instruct", - "mobilint/Qwen2.5-7B-Instruct" - ] + "bezzam/Qwen3-ASR-0.6B", + "hypaai/Qwen3-ASR-0.6B_2026-03-22_04-35-10" + ], + "relevancy_score": 36.9 }, { - "architecture_id": "MobilintQwen3ForCausalLM", - "total_models": 4, + "architecture_id": "CircuitGPTForCausalLM", + "total_models": 1, + "total_downloads": 1434, + "min_param_count": 419124736, "sample_models": [ - "mobilint/Qwen3-4B", - "mobilint/Qwen3-0.6B", - "mobilint/Qwen3-1.7B", - "mobilint/Qwen3-8B" - ] + "openai/circuit-sparsity" + ], + "relevancy_score": 36.7 }, { - "architecture_id": "MobilintExaoneForCausalLM", - "total_models": 4, + "architecture_id": "ParamBharatGenForCausalLM", + "total_models": 3, + "total_downloads": 4524, + "min_param_count": 2860673024, "sample_models": [ - "mobilint/EXAONE-3.5-2.4B-Instruct", - "mobilint/EXAONE-Deep-2.4B", - "mobilint/EXAONE-3.5-7.8B-Instruct", - "mobilint/EXAONE-Deep-7.8B" - ] + "bharatgenai/Param-1-5B", + "bharatgenai/AyurParam", + "bharatgenai/Param-1-2.9B-Instruct" + ], + "relevancy_score": 36.5 }, { "architecture_id": "Qwen2_5_VLForConditionalGeneration", "total_models": 4, + "total_downloads": 15926, + "min_param_count": 5029522432, "sample_models": [ "nvidia/Qwen2.5-VL-7B-Instruct-NVFP4", "OmniSVG/OmniSVG1.1_4B", "OmniSVG/OmniSVG1.1_8B", "OmniSVG/OmniSVG" - ] + ], + "relevancy_score": 35.9 }, { - "architecture_id": "IdeficsForVisionText2Text", - "total_models": 4, + "architecture_id": "DUO", + "total_models": 1, + "total_downloads": 996, + "min_param_count": 169627250, "sample_models": [ - "HuggingFaceM4/idefics-80b-instruct", - "HuggingFaceM4/idefics-9b", - "HuggingFaceM4/idefics-9b-instruct", - "HuggingFaceM4/idefics-80b" - ] + "s-sahoo/duo-distilled" + ], + "relevancy_score": 35.9 }, { - "architecture_id": "LISAForCausalLM", - "total_models": 4, + "architecture_id": "YoutuForCausalLM", + "total_models": 2, + "total_downloads": 4341, + "min_param_count": 1961560064, "sample_models": [ - "xinlai/LISA-13B-llama2-v1", - "xinlai/LISA-7B-v1", - "xinlai/LISA-7B-v1-explanatory", - "xinlai/LISA-13B-llama2-v1-explanatory" - ] + "tencent/Youtu-LLM-2B-Base", + "tencent/Youtu-LLM-2B" + ], + "relevancy_score": 35.8 }, { - "architecture_id": "LLaDAModelLM", - "total_models": 3, + "architecture_id": "Rwkv7ForCausalLM", + "total_models": 1, + "total_downloads": 941, + "min_param_count": 34158592, "sample_models": [ - "GSAI-ML/LLaDA-8B-Instruct", - "GSAI-ML/LLaDA-8B-Base", - "GSAI-ML/LLaDA-1.5" - ] + "admijgjtjtjtjjg/dfdfdf" + ], + "relevancy_score": 35.8 }, { - "architecture_id": "BambaForCausalLM", - "total_models": 3, + "architecture_id": "RubiRLM", + "total_models": 1, + "total_downloads": 928, + "min_param_count": 988446027, "sample_models": [ - "hmellor/tiny-random-BambaForCausalLM", - "ibm-ai-platform/Bamba-9B-v1", - "ibm-ai-platform/Bamba-9B-v2" - ] + "DevHunterAI/RubiRLM-1B-Base" + ], + "relevancy_score": 35.8 }, { - "architecture_id": "InternLMForCausalLM", - "total_models": 3, + "architecture_id": "RavenForCausalLM", + "total_models": 2, + "total_downloads": 4229, + "min_param_count": 1385228288, "sample_models": [ - "internlm/internlm-chat-7b", - "internlm/internlm-20b", - "internlm/internlm-7b" - ] + "tomg-group-umd/huginn-0125", + "smcleish/Recurrent-Llama-3.2-train-recurrence-32" + ], + "relevancy_score": 35.7 }, { - "architecture_id": "Ernie4_5_MoeForCausalLM", - "total_models": 3, + "architecture_id": "PanguEmbeddedForCausalLM", + "total_models": 1, + "total_downloads": 5420, + "min_param_count": 1391497728, "sample_models": [ - "baidu/ERNIE-4.5-21B-A3B-PT", - "baidu/ERNIE-4.5-21B-A3B-Base-PT", - "baidu/ERNIE-4.5-21B-A3B-Thinking" - ] + "FreedomIntelligence/openPangu-Embedded-1B" + ], + "relevancy_score": 35.7 }, { - "architecture_id": "Exaone4ForCausalLM", - "total_models": 3, + "architecture_id": "SoraForSLM", + "total_models": 1, + "total_downloads": 900, + "min_param_count": 450707456, "sample_models": [ - "LGAI-EXAONE/EXAONE-4.0-1.2B", - "LGAI-EXAONE/EXAONE-4.0.1-32B", - "LGAI-EXAONE/EXAONE-4.0-32B" - ] + "Conlanger-LLM-CLEM/Sorie" + ], + "relevancy_score": 35.7 }, { - "architecture_id": "OlmoHybridForCausalLM", - "total_models": 3, + "architecture_id": "HGRNBitForCausalLM", + "total_models": 1, + "total_downloads": 883, + "min_param_count": 374108160, "sample_models": [ - "allenai/Olmo-Hybrid-7B", - "allenai/Olmo-Hybrid-Instruct-DPO-7B", - "allenai/Olmo-Hybrid-Instruct-SFT-7B" - ] + "ridger/MMfreeLM-370M" + ], + "relevancy_score": 35.7 }, { - "architecture_id": "Llama4ForCausalLM", - "total_models": 3, + "architecture_id": "GTLMForCausalLM", + "total_models": 2, + "total_downloads": 3960, + "min_param_count": 2095989760, "sample_models": [ - "trl-internal-testing/tiny-Llama4ForCausalLM", - "pruna-test/test-save-tiny-random-llama4-smashed", - "facebook/MobileLLM-R1.5-360M" - ] + "Madras1/GTLM-1-2B-A350M", + "Madras1/GTLM-1-2B-A350M-fp16" + ], + "relevancy_score": 35.6 }, { - "architecture_id": "BitNetForCausalLM", - "total_models": 3, + "architecture_id": "MiMoForCausalLM", + "total_models": 2, + "total_downloads": 138763, + "min_param_count": 7833409536, "sample_models": [ - "microsoft/bitnet-b1.58-2B-4T", - "microsoft/bitnet-b1.58-2B-4T-bf16", - "iSolver-AI/FEnet" - ] + "XiaomiMiMo/MiMo-7B-Base", + "XiaomiMiMo/MiMo-7B-RL" + ], + "relevancy_score": 35.5 }, { - "architecture_id": "IQuestCoderForCausalLM", - "total_models": 3, + "architecture_id": "DotLMForCausalLM", + "total_models": 1, + "total_downloads": 822, + "min_param_count": 176204544, "sample_models": [ - "IQuestLab/IQuest-Coder-V1-40B-Instruct", - "IQuestLab/IQuest-Coder-V1-7B-Instruct", - "Multilingual-Multimodal-NLP/IndustrialCoder" - ] + "tensorfiend/DotLM-165M" + ], + "relevancy_score": 35.5 }, { - "architecture_id": "XverseForCausalLM", - "total_models": 3, + "architecture_id": "MoshiForConditionalGeneration", + "total_models": 2, + "total_downloads": 133357, + "min_param_count": 7783880545, "sample_models": [ - "xverse/XVERSE-7B-Chat", - "katuni4ka/tiny-random-xverse", - "xverse/XVERSE-13B-256K" - ] + "kmhf/hf-moshiko", + "kmhf/hf-moshika" + ], + "relevancy_score": 35.4 }, { - "architecture_id": "PersimmonForCausalLM", - "total_models": 3, + "architecture_id": "DeltaNetForCausalLM", + "total_models": 1, + "total_downloads": 4602, + "min_param_count": 1365677056, "sample_models": [ - "adept/persimmon-8b-chat", - "adept/persimmon-8b-base", - "pszemraj/perSLIMmon-8b-base" - ] + "fla-hub/delta_net-1.3B-100B" + ], + "relevancy_score": 35.3 }, { - "architecture_id": "RecurrentGemmaForCausalLM", - "total_models": 3, + "architecture_id": "VaultGemmaForCausalLM", + "total_models": 1, + "total_downloads": 4252, + "min_param_count": 1038741120, "sample_models": [ - "google/recurrentgemma-2b", - "google/recurrentgemma-2b-it", - "google/recurrentgemma-9b" - ] + "google/vaultgemma-1b" + ], + "relevancy_score": 35.2 }, { - "architecture_id": "Llama4ForConditionalGeneration", + "architecture_id": "RWKV7ForCausalLM", "total_models": 3, + "total_downloads": 2190, + "min_param_count": 1527404544, + "sample_models": [ + "RWKV/RWKV7-Goose-World3-1.5B-HF", + "fla-hub/rwkv7-1.5B-world", + "RWKV/RWKV7-Goose-World3-2.9B-HF" + ], + "relevancy_score": 34.9 + }, + { + "architecture_id": "Rwkv5ForCausalLM", + "total_models": 2, + "total_downloads": 2897, + "min_param_count": 1577754624, "sample_models": [ - "yujiepan/llama-4-tiny-random", - "RedHatAI/Llama-4-Scout-17B-16E-Instruct-NVFP4", - "RedHatAI/Llama-4-Maverick-17B-128E-Instruct-NVFP4" - ] + "RWKV/rwkv-5-world-3b", + "RWKV/rwkv-5-world-1b5" + ], + "relevancy_score": 34.9 }, { - "architecture_id": "LlavaLlamaModel", - "total_models": 3, + "architecture_id": "MegaForCausalLM", + "total_models": 1, + "total_downloads": 636, + "min_param_count": 126132108, "sample_models": [ - "Efficient-Large-Model/VILA1.5-3b", - "Efficient-Large-Model/NVILA-Lite-8B", - "Efficient-Large-Model/NVILA-8B" - ] + "BEE-spoke-data/mega-ar-126m-4k" + ], + "relevancy_score": 34.9 }, { - "architecture_id": "AraGPT2LMHeadModel", - "total_models": 3, + "architecture_id": "KimiK2ForCausalLM", + "total_models": 1, + "total_downloads": 582, + "min_param_count": 170595012, "sample_models": [ - "QCRI/Fanar-2-Diwan", - "aubmindlab/aragpt2-mega", - "aubmindlab/aragpt2-large" - ] + "hyper-accel/tiny-random-kimi-k2" + ], + "relevancy_score": 34.7 }, { - "architecture_id": "RITAModelForCausalLM", - "total_models": 3, + "architecture_id": "LilleForCausalLM", + "total_models": 1, + "total_downloads": 546, + "min_param_count": 127236768, "sample_models": [ - "lightonai/RITA_s", - "lightonai/RITA_xl", - "lightonai/RITA_l" - ] + "Nikity/lille-130m-instruct" + ], + "relevancy_score": 34.6 }, { - "architecture_id": "NanoChatForCausalLM", - "total_models": 3, + "architecture_id": "LLaDA2MoeModelLM", + "total_models": 5, + "total_downloads": 245413, + "min_param_count": 16255643392, "sample_models": [ - "Twobombs/nanochat-d34-sft-hf", - "pankajmathur/nanochat-d34-sft-hf", - "Nekochu/nanochat-d24" - ] + "inclusionAI/LLaDA2.1-flash", + "inclusionAI/LLaDA2.0-mini", + "inclusionAI/LLaDA2.1-mini", + "inclusionAI/LLaDA2.0-mini-CAP", + "inclusionAI/LLaDA2.0-flash" + ], + "relevancy_score": 34.5 }, { - "architecture_id": "MobileLlamaForCausalLM", - "total_models": 3, + "architecture_id": "GPT2CompetitiveMoE", + "total_models": 1, + "total_downloads": 526, + "min_param_count": 497796864, "sample_models": [ - "mtgv/MobileVLM_V2-1.7B", - "mtgv/MobileVLM_V2-7B", - "mtgv/MobileVLM_V2-3B" - ] + "Fu01978/gpt2-4x124M-competitive-moe" + ], + "relevancy_score": 34.5 }, { - "architecture_id": "ParamBharatGenForCausalLM", - "total_models": 3, + "architecture_id": "BolmoForCausalLM", + "total_models": 2, + "total_downloads": 2002, + "min_param_count": 1468911776, "sample_models": [ - "bharatgenai/Param-1-5B", - "bharatgenai/AyurParam", - "bharatgenai/Param-1-2.9B-Instruct" - ] + "allenai/Bolmo-1B", + "allenai/Bolmo-7B" + ], + "relevancy_score": 34.1 }, { - "architecture_id": "modeling_camelidae.LlamaForCausalLM", - "total_models": 3, + "architecture_id": "MoELLaVAQwen2ForCausalLM", + "total_models": 1, + "total_downloads": 2591, + "min_param_count": 1406119552, "sample_models": [ - "hywu/Camelidae-8x34B", - "hywu/Camelidae-8x7B", - "hywu/Camelidae-8x13B" - ] + "KKHYA/llavaqwen2.5-0.5b-finetune-moe-4e-2k_20260331_194516" + ], + "relevancy_score": 34.1 }, { - "architecture_id": "MptForCausalLM", - "total_models": 3, + "architecture_id": "XCurOSForCausalLM", + "total_models": 1, + "total_downloads": 93299, + "min_param_count": 7615616512, "sample_models": [ - "yujiepan/mpt-tiny-random", - "explosion-testing/mpt-test", - "team-lucid/mptk-1b" - ] + "XCurOS/XCurOS-0.1-8B-Instruct" + ], + "relevancy_score": 34.0 }, { - "architecture_id": "BlueLMForCausalLM", - "total_models": 3, + "architecture_id": "TarsierForConditionalGeneration", + "total_models": 1, + "total_downloads": 86574, + "min_param_count": 7063427072, "sample_models": [ - "vivo-ai/BlueLM-7B-Chat", - "vivo-ai/BlueLM-7B-Base", - "vivo-ai/BlueLM-7B-Chat-32K" - ] + "omni-research/Tarsier-7b" + ], + "relevancy_score": 33.9 }, { - "architecture_id": "LlamaMoEForCausalLM", - "total_models": 3, + "architecture_id": "OlmoHybridForCausalLM", + "total_models": 4, + "total_downloads": 35668, + "min_param_count": 7430870688, "sample_models": [ - "llama-moe/LLaMA-MoE-v1-3_5B-2_8", - "llama-moe/LLaMA-MoE-v1-3_0B-2_16", - "llama-moe/LLaMA-MoE-v1-3_5B-4_16" - ] + "allenai/Olmo-Hybrid-7B", + "allenai/Olmo-Hybrid-Instruct-DPO-7B", + "allenai/Olmo-Hybrid-Instruct-SFT-7B", + "allenai/Olmo-Hybrid-Think-SFT-7B" + ], + "relevancy_score": 33.7 }, { - "architecture_id": "H2OVLChatModel", - "total_models": 2, + "architecture_id": "SongGenMixedForConditionalGeneration", + "total_models": 1, + "total_downloads": 1849, + "min_param_count": 1363657956, "sample_models": [ - "h2oai/h2ovl-mississippi-800m", - "h2oai/h2ovl-mississippi-2b" - ] + "LiuZH-19/SongGen_mixed_pro" + ], + "relevancy_score": 33.3 }, { - "architecture_id": "KimiK25ForConditionalGeneration", + "architecture_id": "ArgonneModel", "total_models": 2, + "total_downloads": 1321, + "min_param_count": 1273807360, "sample_models": [ - "nvidia/Kimi-K2.5-NVFP4", - "Ex0bit/Kimi-K2.5-PRISM-REAP-530B-A32B" - ] + "PursuitOfDataScience/Argonne2.5-base", + "PursuitOfDataScience/Argonne2.5-instruct" + ], + "relevancy_score": 33.2 }, { - "architecture_id": "HCXVisionV2ForCausalLM", - "total_models": 2, + "architecture_id": "HybridQwen3ForCausalLM", + "total_models": 9, + "total_downloads": 7359, + "min_param_count": 8495712960, "sample_models": [ - "naver-hyperclovax/HyperCLOVAX-SEED-Omni-8B", - "naver-hyperclovax/HyperCLOVAX-SEED-Think-32B" - ] + "amazon/GKA-primed-HQwen3-8B-Instruct", + "amazon/Mamba2-primed-HQwen3-8B-Instruct", + "amazon/GDN-primed-HQwen3-8B-Instruct", + "amazon/GDN-primed-HQwen3-32B-Instruct", + "amazon/GKA-primed-HQwen3-32B-Instruct", + "amazon/BMOJOF-primed-HQwen3-8B-Instruct", + "amazon/GKA-primed-HQwen3-8B-Reasoner", + "amazon/GDN-primed-HQwen3-8B-Reasoner", + "amazon/GKA-primed-HQwen3-32B-Reasoner" + ], + "relevancy_score": 33.1 }, { - "architecture_id": "SolarOpenForCausalLM", + "architecture_id": "JetNemotronForCausalLM", "total_models": 2, + "total_downloads": 7740, + "min_param_count": 3960424768, "sample_models": [ - "upstage/Solar-Open-100B", - "nota-ai/Solar-Open-100B-NotaMoEQuant-Int4" - ] + "jet-ai/Jet-Nemotron-2B", + "jet-ai/Jet-Nemotron-4B" + ], + "relevancy_score": 33.1 }, { - "architecture_id": "OpenAIGPTLMHeadModel", - "total_models": 2, + "architecture_id": "OmniASRForConditionalGeneration", + "total_models": 1, + "total_downloads": 1627, + "min_param_count": 1631506944, "sample_models": [ - "openai-community/openai-gpt", - "lgaalves/gpt1" - ] + "bezzam/omniasr-llm-300m-v2" + ], + "relevancy_score": 33.0 }, { - "architecture_id": "MoshiForConditionalGeneration", - "total_models": 2, + "architecture_id": "MobileLLMP1ForCausalLM", + "total_models": 1, + "total_downloads": 1550, + "min_param_count": 1084453120, "sample_models": [ - "kmhf/hf-moshiko", - "kmhf/hf-moshika" - ] + "facebook/MobileLLM-Pro-base" + ], + "relevancy_score": 32.9 }, { - "architecture_id": "SarvamMLAForCausalLM", - "total_models": 2, + "architecture_id": "Rwkv6ForCausalLM", + "total_models": 8, + "total_downloads": 8264, + "min_param_count": 7635746816, "sample_models": [ - "aoxo/sarvam-105b-uncensored", - "sarvamai/sarvam-105b" - ] + "RWKV/v6-Finch-1B6-HF", + "RWKV/v6-Finch-7B-HF", + "RWKV/rwkv-6-world-1b6", + "RWKV/v6-Finch-14B-HF", + "RWKV/v6-Finch-3B-HF", + "RWKV/rwkv-6-world-7b", + "RWKV/rwkv-6-world-3b-v2.1", + "RWKV/rwkv-6-world-3b" + ], + "relevancy_score": 32.8 }, { - "architecture_id": "ReformerModelWithLMHead", - "total_models": 2, + "architecture_id": "GiddForDiffusionLM", + "total_models": 1, + "total_downloads": 1342, + "min_param_count": 2957629440, "sample_models": [ - "google/reformer-crime-and-punishment", - "google/reformer-enwik8" - ] + "dvruette/gidd-unif-3b" + ], + "relevancy_score": 32.6 }, { - "architecture_id": "GPTNeoXJapaneseForCausalLM", - "total_models": 2, + "architecture_id": "Kanana2VecModel", + "total_models": 1, + "total_downloads": 1330, + "min_param_count": 2086979328, "sample_models": [ - "abeja/gpt-neox-japanese-2.7b", - "hf-tiny-model-private/tiny-random-GPTNeoXJapaneseForCausalLM" - ] + "kakaocorp/kanana-nano-2.1b-embedding" + ], + "relevancy_score": 32.6 }, { - "architecture_id": "SarvamMoEForCausalLM", - "total_models": 2, + "architecture_id": "AeroForConditionalGeneration", + "total_models": 1, + "total_downloads": 1271, + "min_param_count": 2416221184, "sample_models": [ - "aoxo/sarvam-30b-uncensored", - "sarvamai/sarvam-30b" - ] + "lmms-lab/Aero-1-Audio" + ], + "relevancy_score": 32.5 }, { - "architecture_id": "MiMoForCausalLM", - "total_models": 2, + "architecture_id": "JAISLMHeadModel", + "total_models": 5, + "total_downloads": 15354, + "min_param_count": 13462730280, "sample_models": [ - "XiaomiMiMo/MiMo-7B-Base", - "XiaomiMiMo/MiMo-7B-RL" - ] + "inceptionai/jais-13b-chat", + "katuni4ka/tiny-random-jais", + "inceptionai/jais-family-30b-8k", + "inceptionai/jais-13b", + "inceptionai/jais-family-13b-chat" + ], + "relevancy_score": 32.4 }, { - "architecture_id": "StarVectorForCausalLM", - "total_models": 2, + "architecture_id": "StableLMAlphaForCausalLM", + "total_models": 1, + "total_downloads": 7278, + "min_param_count": 6889414656, "sample_models": [ - "starvector/starvector-1b-im2svg", - "starvector/starvector-8b-im2svg" - ] + "stabilityai/stablelm-base-alpha-7b-v2" + ], + "relevancy_score": 32.4 }, { - "architecture_id": "BaiChuanForCausalLM", - "total_models": 2, + "architecture_id": "DeciCoderForCausalLM", + "total_models": 1, + "total_downloads": 1227, + "min_param_count": 1113671680, "sample_models": [ - "baichuan-inc/Baichuan-7B", - "FreedomIntelligence/HuatuoGPT-7B" - ] + "Deci/DeciCoder-1b" + ], + "relevancy_score": 32.4 }, { - "architecture_id": "MiniMaxM1ForCausalLM", - "total_models": 2, + "architecture_id": "XLNetLMHeadModel", + "total_models": 5, + "total_downloads": 489758, + "min_param_count": null, "sample_models": [ - "MiniMaxAI/MiniMax-M1-40k", - "MiniMaxAI/MiniMax-M1-80k" - ] + "xlnet/xlnet-base-cased", + "xlnet/xlnet-large-cased", + "hfl/chinese-xlnet-base", + "sshleifer/tiny-xlnet-base-cased", + "textattack/xlnet-base-cased-imdb" + ], + "relevancy_score": 32.1 }, { - "architecture_id": "DeepseekForCausalLM", - "total_models": 2, + "architecture_id": "IQuestCoderForCausalLM", + "total_models": 4, + "total_downloads": 17787, + "min_param_count": 7612810240, "sample_models": [ - "deepseek-ai/deepseek-moe-16b-base", - "deepseek-ai/deepseek-moe-16b-chat" - ] + "IQuestLab/IQuest-Coder-V1-40B-Instruct", + "IQuestLab/IQuest-Coder-V1-7B-Instruct", + "Multilingual-Multimodal-NLP/IndustrialCoder", + "IQuestLab/IQuest-Coder-V1-40B-Thinking" + ], + "relevancy_score": 32.1 }, { - "architecture_id": "Phi3SmallForCausalLM", - "total_models": 2, + "architecture_id": "GritLM", + "total_models": 1, + "total_downloads": 29668, + "min_param_count": 7241732096, "sample_models": [ - "microsoft/Phi-3-small-8k-instruct", - "microsoft/Phi-3-small-128k-instruct" - ] + "parasail-ai/GritLM-7B-vllm" + ], + "relevancy_score": 31.5 }, { - "architecture_id": "EchoForCausalLM", + "architecture_id": "KimiK25ForConditionalGeneration", "total_models": 2, + "total_downloads": 799324, + "min_param_count": 91383180528, "sample_models": [ - "ethicalabs/Echo-DSRN-486M-v0.7.6-SFT", - "ethicalabs/Echo-DSRN-114M-Base" - ] + "nvidia/Kimi-K2.5-NVFP4", + "Ex0bit/Kimi-K2.5-PRISM-REAP-530B-A32B" + ], + "relevancy_score": 31.4 }, { - "architecture_id": "Ernie4_5ForCausalLM", - "total_models": 2, + "architecture_id": "CambrianQwenForCausalLM", + "total_models": 1, + "total_downloads": 4573, + "min_param_count": 3986951616, "sample_models": [ - "baidu/ERNIE-4.5-0.3B-PT", - "baidu/ERNIE-4.5-0.3B-Base-PT" - ] + "nyu-visionx/Scale-RAE-Qwen1.5B_DiT2.4B" + ], + "relevancy_score": 31.3 }, { - "architecture_id": "OrionForCausalLM", - "total_models": 2, + "architecture_id": "Lfm2Prototype1ForCausalLM", + "total_models": 1, + "total_downloads": 733, + "min_param_count": 1212304128, "sample_models": [ - "OrionStarAI/Orion-14B-Chat", - "OrionStarAI/Orion-14B-Base" - ] + "nntsuzu/LFM2-SFT-Prototype01-1.2B-JP" + ], + "relevancy_score": 31.3 }, { - "architecture_id": "NemotronFlashForCausalLM", - "total_models": 2, + "architecture_id": "InternVLChatModel", + "total_models": 1, + "total_downloads": 4303, + "min_param_count": 3712637952, "sample_models": [ - "nvidia/Nemotron-Flash-3B", - "nvidia/Nemotron-Flash-1B" - ] + "numind/NuExtract-2-4B-experimental" + ], + "relevancy_score": 31.2 }, { "architecture_id": "AXK1ForCausalLM", "total_models": 2, + "total_downloads": 19063, + "min_param_count": 11448603648, "sample_models": [ "skt/A.X-K1", "thkim93/axk1-2layers" - ] + ], + "relevancy_score": 31.1 }, { - "architecture_id": "DbrxForCausalLM", - "total_models": 2, + "architecture_id": "CheXagentForCausalLM", + "total_models": 1, + "total_downloads": 4115, + "min_param_count": 3140746752, + "sample_models": [ + "StanfordAIMI/CheXagent-2-3b" + ], + "relevancy_score": 31.1 + }, + { + "architecture_id": "HymbaForCausalLM", + "total_models": 1, + "total_downloads": 682, + "min_param_count": 1522797824, + "sample_models": [ + "nvidia/Hymba-1.5B-Instruct" + ], + "relevancy_score": 31.1 + }, + { + "architecture_id": "IdeficsForVisionText2Text", + "total_models": 4, + "total_downloads": 10253, + "min_param_count": 8929682192, "sample_models": [ - "trl-internal-testing/tiny-DbrxForCausalLM", - "katuni4ka/tiny-random-dbrx" - ] + "HuggingFaceM4/idefics-80b-instruct", + "HuggingFaceM4/idefics-9b", + "HuggingFaceM4/idefics-9b-instruct", + "HuggingFaceM4/idefics-80b" + ], + "relevancy_score": 30.9 }, { - "architecture_id": "Dots1ForCausalLM", - "total_models": 2, + "architecture_id": "PersimmonForCausalLM", + "total_models": 3, + "total_downloads": 13281, + "min_param_count": 8823735296, "sample_models": [ - "rednote-hilab/dots.llm1.inst", - "rednote-hilab/dots.llm1.base" - ] + "adept/persimmon-8b-chat", + "adept/persimmon-8b-base", + "pszemraj/perSLIMmon-8b-base" + ], + "relevancy_score": 30.9 }, { - "architecture_id": "FlexOlmoForCausalLM", + "architecture_id": "Phi3SmallForCausalLM", "total_models": 2, + "total_downloads": 17781, + "min_param_count": 7392272384, "sample_models": [ - "allenai/Flex-reddit-2x7B-1T", - "allenai/FlexOlmo-7x7B-1T-RT" - ] + "microsoft/Phi-3-small-8k-instruct", + "microsoft/Phi-3-small-128k-instruct" + ], + "relevancy_score": 30.9 }, { "architecture_id": "ChatGLMModel", "total_models": 2, + "total_downloads": 17530, + "min_param_count": 9399951392, "sample_models": [ "zai-org/codegeex4-all-9b", "zai-org/glm-4-9b" - ] + ], + "relevancy_score": 30.9 }, { - "architecture_id": "CLIPT5ForConditionalGeneration", - "total_models": 2, + "architecture_id": "FlexOlmoForCausalLM", + "total_models": 3, + "total_downloads": 12636, + "min_param_count": 11627401216, "sample_models": [ - "zhiqiulin/clip-flant5-xl", - "zhiqiulin/clip-flant5-xxl" - ] + "allenai/Flex-reddit-2x7B-1T", + "allenai/FlexOlmo-7x7B-1T-RT", + "shanearora/Flex-reddit-2x7B-1T" + ], + "relevancy_score": 30.8 }, { - "architecture_id": "PenguinVLQwen3ForCausalLM", - "total_models": 2, + "architecture_id": "SpatialLMLlamaForCausalLM", + "total_models": 1, + "total_downloads": 580, + "min_param_count": 1345883776, "sample_models": [ - "tencent/Penguin-VL-8B", - "tencent/Penguin-VL-2B" - ] + "manycore-research/SpatialLM1.1-Llama-1B" + ], + "relevancy_score": 30.7 + }, + { + "architecture_id": "MixFormerSequentialForCausalLM", + "total_models": 1, + "total_downloads": 543, + "min_param_count": 2779683840, + "sample_models": [ + "SkunkworksAI/phi-2" + ], + "relevancy_score": 30.6 }, { "architecture_id": "StripedHyenaModelForCausalLM", - "total_models": 2, + "total_models": 3, + "total_downloads": 10737, + "min_param_count": 7646024704, "sample_models": [ "togethercomputer/evo-1-131k-base", - "togethercomputer/evo-1-8k-base" - ] + "togethercomputer/evo-1-8k-base", + "togethercomputer/StripedHyena-Nous-7B" + ], + "relevancy_score": 30.4 }, { - "architecture_id": "CrystalCoderLMHeadModel", - "total_models": 2, + "architecture_id": "Maira2ForConditionalGeneration", + "total_models": 1, + "total_downloads": 2817, + "min_param_count": 6880185600, "sample_models": [ - "LLM360/Crystal", - "LLM360/CrystalChat" - ] + "microsoft/maira-2" + ], + "relevancy_score": 30.2 }, { - "architecture_id": "JetNemotronForCausalLM", - "total_models": 2, + "architecture_id": "BioGptForCausalLM", + "total_models": 5, + "total_downloads": 189545, + "min_param_count": null, "sample_models": [ - "jet-ai/Jet-Nemotron-2B", - "jet-ai/Jet-Nemotron-4B" - ] + "microsoft/biogpt", + "microsoft/BioGPT-Large", + "microsoft/BioGPT-Large-PubMedQA", + "hf-tiny-model-private/tiny-random-BioGptForCausalLM", + "zequnl/molxpt" + ], + "relevancy_score": 30.0 }, { - "architecture_id": "Mamba2ForCausalLM", - "total_models": 2, + "architecture_id": "Ernie4_5_MoeForCausalLM", + "total_models": 4, + "total_downloads": 39656, + "min_param_count": 21825437888, "sample_models": [ - "deqing/mamba2-300M-v5-mamba2", - "EchoLabs33/mamba2-1.3b-hxq" - ] + "baidu/ERNIE-4.5-21B-A3B-PT", + "baidu/ERNIE-4.5-21B-A3B-Base-PT", + "baidu/ERNIE-4.5-21B-A3B-Thinking", + "baidu/ERNIE-4.5-300B-A47B-PT" + ], + "relevancy_score": 29.9 }, { - "architecture_id": "MolformerForCausalLM", - "total_models": 2, + "architecture_id": "SeedOssForCausalLM", + "total_models": 4, + "total_downloads": 34422, + "min_param_count": 21030639488, "sample_models": [ - "ibm-research/GP-MoLFormer-Uniq", - "ralyn/NPComposer-v2" - ] + "ByteDance-Seed/Seed-OSS-36B-Instruct", + "NousResearch/Hermes-4.3-36B", + "ByteDance-Seed/Seed-OSS-36B-Base", + "mratsim/Seed-OSS-36B-Instruct-NVFP4" + ], + "relevancy_score": 29.6 }, { - "architecture_id": "CogVLMForCausalLM", - "total_models": 2, + "architecture_id": "BailingMoeV2ForCausalLM", + "total_models": 5, + "total_downloads": 20877, + "min_param_count": 16255643392, "sample_models": [ - "zai-org/cogvlm2-llama3-chat-19B", - "zai-org/cogvlm-chat-hf" - ] + "inclusionAI/Ling-mini-2.0", + "inclusionAI/Ling-1T", + "inclusionAI/Ring-mini-2.0", + "inclusionAI/Ling-flash-2.0", + "inclusionAI/Ling-flash-base-2.0" + ], + "relevancy_score": 29.1 }, { - "architecture_id": "Jais2ForCausalLM", - "total_models": 2, + "architecture_id": "LatentMoELLaVAPhiForCausalLM", + "total_models": 1, + "total_downloads": 1700, + "min_param_count": 3093139456, "sample_models": [ - "inceptionai/Jais-2-8B-Chat", - "inceptionai/Jais-2-70B-Chat" - ] + "KKHYA/llavaphi2-2.7b-finetune-latent-sparse-moe-4e-2k-freeze-1.0_20260304_075653" + ], + "relevancy_score": 29.1 }, { "architecture_id": "Qwen2ForSequenceClassification", "total_models": 2, + "total_downloads": 7018, + "min_param_count": 7070622720, "sample_models": [ "nvidia/AceMath-7B-RM", "nvidia/Qwen2.5-CascadeRL-RM-72B" - ] + ], + "relevancy_score": 28.9 }, { - "architecture_id": "ChatGLMForConditionalGeneration", - "total_models": 2, + "architecture_id": "MatriochkaForCausalLM", + "total_models": 1, + "total_downloads": 1557, + "min_param_count": 3358735360, "sample_models": [ - "IAAR-Shanghai/xVerify-9B-C", - "qiuhuachuan/MeChat" - ] + "nthngdy/matryoshka-3B" + ], + "relevancy_score": 28.9 }, { - "architecture_id": "RavenForCausalLM", + "architecture_id": "SolarOpenForCausalLM", "total_models": 2, + "total_downloads": 244402, + "min_param_count": null, "sample_models": [ - "tomg-group-umd/huginn-0125", - "smcleish/Recurrent-Llama-3.2-train-recurrence-32" - ] + "upstage/Solar-Open-100B", + "nota-ai/Solar-Open-100B-NotaMoEQuant-Int4" + ], + "relevancy_score": 28.8 }, { - "architecture_id": "YoutuForCausalLM", + "architecture_id": "DeepseekForCausalLM", "total_models": 2, + "total_downloads": 41860, + "min_param_count": 16375728128, "sample_models": [ - "tencent/Youtu-LLM-2B-Base", - "tencent/Youtu-LLM-2B" - ] + "deepseek-ai/deepseek-moe-16b-base", + "deepseek-ai/deepseek-moe-16b-chat" + ], + "relevancy_score": 28.8 }, { - "architecture_id": "Qwen3VLMoeForConditionalGeneration", + "architecture_id": "Jais2ForCausalLM", "total_models": 2, + "total_downloads": 6234, + "min_param_count": 8090401280, "sample_models": [ - "RedHatAI/Qwen3-VL-235B-A22B-Instruct-NVFP4", - "Oysiyl/qwen3-vl-30b-a3b-unslop-good-lora-v1" - ] + "inceptionai/Jais-2-8B-Chat", + "inceptionai/Jais-2-70B-Chat" + ], + "relevancy_score": 28.6 }, { - "architecture_id": "MosaicGPT", - "total_models": 2, + "architecture_id": "HyperCLOVAXForCausalLM", + "total_models": 1, + "total_downloads": 43175, + "min_param_count": 14748112896, "sample_models": [ - "anas-awadalla/mpt-1b-redpajama-200b", - "anas-awadalla/mpt-1b-redpajama-200b-dolly" - ] + "naver-hyperclovax/HyperCLOVAX-SEED-Think-14B" + ], + "relevancy_score": 28.3 }, { - "architecture_id": "GTLMForCausalLM", - "total_models": 2, + "architecture_id": "LamedPhi3ForCausalLM", + "total_models": 1, + "total_downloads": 1181, + "min_param_count": 4049101904, "sample_models": [ - "Madras1/GTLM-1-2B-A350M", - "Madras1/GTLM-1-2B-A350M-fp16" - ] + "GoodBaiBai88/M3D-LaMed-Phi-3-4B" + ], + "relevancy_score": 28.3 }, { - "architecture_id": "GPT2Model", + "architecture_id": "ChatGLMForConditionalGeneration", "total_models": 2, + "total_downloads": 4856, + "min_param_count": 9399951392, "sample_models": [ - "cerebras/Cerebras-GPT-13B", - "keshan/sinhala-gpt2" - ] + "IAAR-Shanghai/xVerify-9B-C", + "qiuhuachuan/MeChat" + ], + "relevancy_score": 28.0 }, { - "architecture_id": "LiquidForCausalLM", + "architecture_id": "ReformerModelWithLMHead", "total_models": 2, + "total_downloads": 159316, + "min_param_count": null, "sample_models": [ - "reaperdoesntknow/DNA-175M", - "reaperdoesntknow/DNA-50M" - ] + "google/reformer-crime-and-punishment", + "google/reformer-enwik8" + ], + "relevancy_score": 27.8 }, { - "architecture_id": "ModernBertDecoderForCausalLM", + "architecture_id": "SarvamMLAForCausalLM", "total_models": 2, + "total_downloads": 150715, + "min_param_count": 55732545631, "sample_models": [ - "jhu-clsp/ettin-decoder-400m", - "jhu-clsp/ettin-decoder-32m" - ] + "aoxo/sarvam-105b-uncensored", + "sarvamai/sarvam-105b" + ], + "relevancy_score": 27.7 }, { - "architecture_id": "GLAForCausalLM", + "architecture_id": "SarvamMoEForCausalLM", "total_models": 2, + "total_downloads": 150615, + "min_param_count": 32152650368, "sample_models": [ - "fla-hub/gla-340M-15B", - "fla-hub/gla-1.3B-100B" - ] + "aoxo/sarvam-30b-uncensored", + "sarvamai/sarvam-30b" + ], + "relevancy_score": 27.7 }, { - "architecture_id": "DuchifatCore", + "architecture_id": "WeDLMForCausalLM", "total_models": 2, + "total_downloads": 3944, + "min_param_count": 8190735360, "sample_models": [ - "Raziel1234/Duchifat-2", - "razielAI/Duchifat-2.1-Instruct" - ] + "tencent/WeDLM-8B-Base", + "tencent/WeDLM-8B-Instruct" + ], + "relevancy_score": 27.6 }, { - "architecture_id": "KonkanGPT", - "total_models": 2, + "architecture_id": "InternLMForCausalLM", + "total_models": 4, + "total_downloads": 80000, + "min_param_count": null, "sample_models": [ - "omdeep22/Gonyai-teo2", - "omdeep22/Gonyai-v1" - ] + "internlm/internlm-chat-7b", + "internlm/internlm-20b", + "internlm/internlm-7b", + "internlm/internlm-chat-20b" + ], + "relevancy_score": 27.5 }, { - "architecture_id": "BertLMHeadModel", - "total_models": 2, + "architecture_id": "SparseLlamaForCausalLM", + "total_models": 1, + "total_downloads": 4652, + "min_param_count": 8185270336, "sample_models": [ - "dicta-il/BEREL_3.0", - "hf-tiny-model-private/tiny-random-BertLMHeadModel" - ] + "openbmb/NOSA-8B" + ], + "relevancy_score": 27.4 }, { - "architecture_id": "RobertaForCausalLM", - "total_models": 2, + "architecture_id": "LongLlamaForCausalLM", + "total_models": 1, + "total_downloads": 746, + "min_param_count": 3426474900, "sample_models": [ - "gokceuludogan/ChemBERTaLM", - "uf-aice-lab/math-roberta" - ] + "syzymon/long_llama_3b" + ], + "relevancy_score": 27.3 }, { - "architecture_id": "MossForCausalLM", + "architecture_id": "Step3p5ForCausalLM", "total_models": 2, + "total_downloads": 119373, + "min_param_count": 111271709328, "sample_models": [ - "OpenMOSS-Team/moss-moon-003-sft", - "OpenMOSS-Team/moss-moon-003-base" - ] + "stepfun-ai/Step-3.5-Flash", + "tacos4me/Step-3.5-Flash-NVFP4" + ], + "relevancy_score": 27.2 }, { - "architecture_id": "WeDLMForCausalLM", + "architecture_id": "GPTNeoXJapaneseForCausalLM", "total_models": 2, + "total_downloads": 113295, + "min_param_count": null, "sample_models": [ - "tencent/WeDLM-8B-Base", - "tencent/WeDLM-8B-Instruct" - ] + "abeja/gpt-neox-japanese-2.7b", + "hf-tiny-model-private/tiny-random-GPTNeoXJapaneseForCausalLM" + ], + "relevancy_score": 27.1 }, { - "architecture_id": "Rwkv5ForCausalLM", - "total_models": 2, + "architecture_id": "ZambaForCausalLM", + "total_models": 1, + "total_downloads": 4137, + "min_param_count": 7232490496, "sample_models": [ - "RWKV/rwkv-5-world-3b", - "RWKV/rwkv-5-world-1b5" - ] + "Zyphra/Zamba-7B-v1" + ], + "relevancy_score": 27.1 }, { - "architecture_id": "BartForCausalLM", - "total_models": 2, + "architecture_id": "JetMoEForCausalLM", + "total_models": 1, + "total_downloads": 3698, + "min_param_count": 8522237952, "sample_models": [ - "sanchit-gandhi/tiny-random-bart-fp16", - "hf-tiny-model-private/tiny-random-BartForCausalLM" - ] + "jetmoe/jetmoe-8b" + ], + "relevancy_score": 26.9 }, { - "architecture_id": "BitnetForCausalLM", - "total_models": 2, + "architecture_id": "Gemma4TextModel", + "total_models": 1, + "total_downloads": 635, + "min_param_count": 4647449856, "sample_models": [ - "1bitLLM/bitnet_b1_58-large", - "1bitLLM/bitnet_b1_58-3B" - ] + "bRadu/gemma-4-E2B-it-textonly" + ], + "relevancy_score": 26.9 }, { - "architecture_id": "Int8OPTForCausalLM", - "total_models": 2, + "architecture_id": "BailingMoeForCausalLM", + "total_models": 1, + "total_downloads": 22008, + "min_param_count": 16801974272, "sample_models": [ - "mit-han-lab/opt-125m-smoothquant", - "mit-han-lab/opt-6.7b-smoothquant" - ] + "inclusionAI/Ling-lite-1.5" + ], + "relevancy_score": 26.8 }, { - "architecture_id": "Olmo2ForSequenceClassification", - "total_models": 2, + "architecture_id": "CXRMate2ForConditionalGeneration", + "total_models": 1, + "total_downloads": 557, + "min_param_count": 3322260224, "sample_models": [ - "allenai/OLMo-2-1124-7B-RM", - "LifeWiki-ai/OLMo-2-1124-7B-RM" - ] + "aehrc/cxrmate-2" + ], + "relevancy_score": 26.6 }, { - "architecture_id": "TranceptionLMHeadModel", - "total_models": 2, + "architecture_id": "BunnyPhiForCausalLM", + "total_models": 1, + "total_downloads": 502, + "min_param_count": 3182254624, "sample_models": [ - "PascalNotin/Tranception_Large", - "PascalNotin/Tranception_Small" - ] + "BAAI/Bunny-v1_0-3B" + ], + "relevancy_score": 26.4 }, { - "architecture_id": "MultiScaleForCausalLM", + "architecture_id": "Esm2LlamaInstructForCausalLM", + "total_models": 1, + "total_downloads": 2601, + "min_param_count": 10878983201, + "sample_models": [ + "xiao-fei/Prot2Text-V2-11B-Instruct-hf" + ], + "relevancy_score": 26.1 + }, + { + "architecture_id": "OuroForCausalLM", + "total_models": 4, + "total_downloads": 34551, + "min_param_count": null, + "sample_models": [ + "ByteDance/Ouro-1.4B", + "ByteDance/Ouro-2.6B-Thinking", + "ByteDance/Ouro-2.6B", + "ByteDance/Ouro-1.4B-Thinking" + ], + "relevancy_score": 25.6 + }, + { + "architecture_id": "StableDiffcoderForCausalLM", "total_models": 2, + "total_downloads": 1641, + "min_param_count": 8250462208, "sample_models": [ - "KoinicLabs/AXL-Vision-v2", - "KoinicLabs/AXL-Translate" - ] + "ByteDance-Seed/Stable-DiffCoder-8B-Instruct", + "ByteDance-Seed/Stable-DiffCoder-8B-Base" + ], + "relevancy_score": 25.6 }, { - "architecture_id": "GPT", + "architecture_id": "BaiChuanForCausalLM", "total_models": 2, + "total_downloads": 55785, + "min_param_count": null, "sample_models": [ - "LH-Tech-AI/Apex-1.5-Coder-Instruct-350M", - "LH-Tech-AI/Apex-1.5-Instruct-350M" - ] + "baichuan-inc/Baichuan-7B", + "FreedomIntelligence/HuatuoGPT-7B" + ], + "relevancy_score": 25.5 + }, + { + "architecture_id": "FP8Qwen3ForCausalLM", + "total_models": 1, + "total_downloads": 1941, + "min_param_count": 8190735360, + "sample_models": [ + "xihc-ucb/Qwen3-8B-Base-train-Quasar-0809" + ], + "relevancy_score": 25.4 }, { - "architecture_id": "BolmoForCausalLM", - "total_models": 2, + "architecture_id": "Qwen2VLAudioForConditionalGeneration", + "total_models": 1, + "total_downloads": 1806, + "min_param_count": 8932935680, "sample_models": [ - "allenai/Bolmo-1B", - "allenai/Bolmo-7B" - ] + "MayaKD/qwen2-vl-audio" + ], + "relevancy_score": 25.3 }, { - "architecture_id": "OpenMoeForCausalLM", - "total_models": 2, + "architecture_id": "HCXVisionForCausalLM", + "total_models": 1, + "total_downloads": 65558, + "min_param_count": null, "sample_models": [ - "hpcai-tech/openmoe-8B", - "OrionZheng/openmoe-8b" - ] + "naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B" + ], + "relevancy_score": 25.2 }, { - "architecture_id": "MiniMindForCausalLM", - "total_models": 2, + "architecture_id": "FP8Qwen2ForCausalLM", + "total_models": 1, + "total_downloads": 1774, + "min_param_count": 7615616512, "sample_models": [ - "yiwenX/MiniMind-MoE-640-120M", - "chujiamo/baiheng_0405" - ] + "xihc-ucb/Qwen2.5-7B-train-Quasar-1214" + ], + "relevancy_score": 25.2 }, { - "architecture_id": "RWKV7ForCausalLM", - "total_models": 2, + "architecture_id": "KimiLinearForCausalLM", + "total_models": 1, + "total_downloads": 61633, + "min_param_count": 49122681728, "sample_models": [ - "RWKV/RWKV7-Goose-World3-1.5B-HF", - "fla-hub/rwkv7-1.5B-world" - ] + "moonshotai/Kimi-Linear-48B-A3B-Instruct" + ], + "relevancy_score": 25.1 }, { - "architecture_id": "BottleneckT5LMWithPerturb", - "total_models": 2, + "architecture_id": "CheXagentForConditionalGeneration", + "total_models": 1, + "total_downloads": 1714, + "min_param_count": 8362401664, "sample_models": [ - "thesephist/contra-bottleneck-t5-small-wikipedia", - "thesephist/contra-bottleneck-t5-base-wikipedia" - ] + "StanfordAIMI/CheXagent-8b" + ], + "relevancy_score": 25.1 }, { - "architecture_id": "StableDiffcoderForCausalLM", - "total_models": 2, + "architecture_id": "MiMoV2FlashForCausalLM", + "total_models": 1, + "total_downloads": 58502, + "min_param_count": 309785318400, "sample_models": [ - "ByteDance-Seed/Stable-DiffCoder-8B-Instruct", - "ByteDance-Seed/Stable-DiffCoder-8B-Base" - ] + "XiaomiMiMo/MiMo-V2-Flash" + ], + "relevancy_score": 25.0 }, { - "architecture_id": "OtterForConditionalGeneration", - "total_models": 2, + "architecture_id": "MobilintLlamaForCausalLM", + "total_models": 4, + "total_downloads": 24984, + "min_param_count": null, "sample_models": [ - "luodian/OTTER-Video-LLaMA7B-DenseCaption", - "luodian/OTTER-MPT1B-RPJama-Init" - ] + "mobilint/Llama-3.2-3B-Instruct", + "mobilint/Llama-3.2-1B-Instruct", + "mobilint/Llama-3.1-8B-Instruct", + "mobilint/HyperCLOVAX-SEED-Text-Instruct-1.5B" + ], + "relevancy_score": 24.9 }, { - "architecture_id": "MonkeyLMHeadModel", - "total_models": 2, + "architecture_id": "MobilintQwen2ForCausalLM", + "total_models": 4, + "total_downloads": 24843, + "min_param_count": null, "sample_models": [ - "echo840/Monkey-Chat", - "echo840/Monkey" - ] + "mobilint/Qwen2.5-1.5B-Instruct", + "mobilint/Qwen2.5-0.5B-Instruct", + "mobilint/Qwen2.5-3B-Instruct", + "mobilint/Qwen2.5-7B-Instruct" + ], + "relevancy_score": 24.9 }, { - "architecture_id": "IndexForCausalLM", - "total_models": 2, + "architecture_id": "MobilintQwen3ForCausalLM", + "total_models": 4, + "total_downloads": 24000, + "min_param_count": null, "sample_models": [ - "IndexTeam/Index-1.9B-Chat", - "IndexTeam/Index-1.9B-Pure" - ] + "mobilint/Qwen3-4B", + "mobilint/Qwen3-0.6B", + "mobilint/Qwen3-1.7B", + "mobilint/Qwen3-8B" + ], + "relevancy_score": 24.8 }, { - "architecture_id": "PointLLMLlamaForCausalLM", - "total_models": 2, + "architecture_id": "KORMoForCausalLM", + "total_models": 1, + "total_downloads": 1376, + "min_param_count": 10756624384, "sample_models": [ - "RunsenXu/PointLLM_7B_v1.1_init", - "RunsenXu/PointLLM_7B_v1.2" - ] + "KORMo-Team/KORMo-10B-sft" + ], + "relevancy_score": 24.7 }, { - "architecture_id": "T5EncoderModel", + "architecture_id": "MiniCPMSALAForCausalLM", "total_models": 1, + "total_downloads": 1359, + "min_param_count": 9477203968, "sample_models": [ - "XLabs-AI/xflux_text_encoders" - ] + "openbmb/MiniCPM-SALA" + ], + "relevancy_score": 24.6 }, { - "architecture_id": "Step3p5ForCausalLM", + "architecture_id": "LongcatFlashForCausalLM", "total_models": 1, + "total_downloads": 46236, + "min_param_count": 561862880256, "sample_models": [ - "stepfun-ai/Step-3.5-Flash" - ] + "meituan-longcat/LongCat-Flash-Chat" + ], + "relevancy_score": 24.5 }, { - "architecture_id": "AprielForCausalLM", + "architecture_id": "TrillionForCausalLM", "total_models": 1, + "total_downloads": 7530, + "min_param_count": 20725519360, "sample_models": [ - "ServiceNow-AI/Apriel-5B-Instruct" - ] + "trillionlabs/Tri-21B-Think" + ], + "relevancy_score": 24.4 }, { - "architecture_id": "IlamaForCausalLM", - "total_models": 1, + "architecture_id": "MobilintExaoneForCausalLM", + "total_models": 4, + "total_downloads": 19054, + "min_param_count": null, "sample_models": [ - "hmellor/Ilama-3.2-1B" - ] + "mobilint/EXAONE-3.5-2.4B-Instruct", + "mobilint/EXAONE-Deep-2.4B", + "mobilint/EXAONE-3.5-7.8B-Instruct", + "mobilint/EXAONE-Deep-7.8B" + ], + "relevancy_score": 24.3 }, { - "architecture_id": "XCurOSForCausalLM", + "architecture_id": "InternLM3ForCausalLM", "total_models": 1, + "total_downloads": 43602, + "min_param_count": null, "sample_models": [ - "XCurOS/XCurOS-0.1-8B-Instruct" - ] + "internlm/internlm3-8b-instruct" + ], + "relevancy_score": 24.3 }, { - "architecture_id": "TarsierForConditionalGeneration", + "architecture_id": "Emu3ForCausalLM", "total_models": 1, + "total_downloads": 1165, + "min_param_count": 8492011520, "sample_models": [ - "omni-research/Tarsier-7b" - ] + "BAAI/Emu3-Chat" + ], + "relevancy_score": 24.3 }, { - "architecture_id": "Plamo2ForCausalLM", + "architecture_id": "BunnyLlamaForCausalLM", "total_models": 1, + "total_downloads": 1149, + "min_param_count": 8479990848, "sample_models": [ - "pfnet/plamo-2-1b" - ] + "typhoon-ai/llama-3-typhoon-v1.5-8b-vision-preview" + ], + "relevancy_score": 24.3 }, { - "architecture_id": "HCXVisionForCausalLM", - "total_models": 1, + "architecture_id": "CogVLMForCausalLM", + "total_models": 2, + "total_downloads": 5273, + "min_param_count": 17639687424, "sample_models": [ - "naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B" - ] + "zai-org/cogvlm2-llama3-chat-19B", + "zai-org/cogvlm-chat-hf" + ], + "relevancy_score": 24.2 }, { - "architecture_id": "KimiLinearForCausalLM", + "architecture_id": "Param2MoEForCausalLM", "total_models": 1, + "total_downloads": 6504, + "min_param_count": 17151140480, "sample_models": [ - "moonshotai/Kimi-Linear-48B-A3B-Instruct" - ] + "bharatgenai/Param2-17B-A2.4B-Thinking" + ], + "relevancy_score": 24.1 }, { - "architecture_id": "MiMoV2FlashForCausalLM", + "architecture_id": "SteerlingForCausalLM", "total_models": 1, + "total_downloads": 1033, + "min_param_count": 8391778304, "sample_models": [ - "XiaomiMiMo/MiMo-V2-Flash" - ] + "guidelabs/steerling-8b" + ], + "relevancy_score": 24.0 }, { - "architecture_id": "LongcatFlashForCausalLM", + "architecture_id": "ICONNForCausalLM", "total_models": 1, + "total_downloads": 938, + "min_param_count": 7833409536, "sample_models": [ - "meituan-longcat/LongCat-Flash-Chat" - ] + "ICONNAI/ICONN-1-Mini-Beta" + ], + "relevancy_score": 23.8 }, { - "architecture_id": "InternLM3ForCausalLM", - "total_models": 1, + "architecture_id": "MiniMaxM1ForCausalLM", + "total_models": 2, + "total_downloads": 24599, + "min_param_count": null, "sample_models": [ - "internlm/internlm3-8b-instruct" - ] + "MiniMaxAI/MiniMax-M1-40k", + "MiniMaxAI/MiniMax-M1-80k" + ], + "relevancy_score": 23.7 }, { - "architecture_id": "HyperCLOVAXForCausalLM", + "architecture_id": "Qwen2VLForConditionalGeneration", "total_models": 1, + "total_downloads": 772, + "min_param_count": 8291375616, "sample_models": [ - "naver-hyperclovax/HyperCLOVAX-SEED-Think-14B" - ] + "typhoon-ai/typhoon2-qwen2vl-7b-vision-instruct" + ], + "relevancy_score": 23.4 }, { - "architecture_id": "GritLM", + "architecture_id": "CogVLMVideoForCausalLM", "total_models": 1, + "total_downloads": 728, + "min_param_count": 12507532544, "sample_models": [ - "parasail-ai/GritLM-7B-vllm" - ] + "zai-org/VisionReward-Video" + ], + "relevancy_score": 23.2 }, { "architecture_id": "BailingMoeV2_5ForCausalLM", "total_models": 1, + "total_downloads": 24973, + "min_param_count": 1012474606720, "sample_models": [ "inclusionAI/Ring-2.5-1T" - ] + ], + "relevancy_score": 23.1 }, { "architecture_id": "ExaoneMoEForCausalLM", "total_models": 1, + "total_downloads": 24188, + "min_param_count": 237099669632, "sample_models": [ "LGAI-EXAONE/K-EXAONE-236B-A23B" - ] + ], + "relevancy_score": 23.0 }, { - "architecture_id": "Grok1ModelForCausalLM", + "architecture_id": "Ernie4_5ForCausalLM", + "total_models": 2, + "total_downloads": 17641, + "min_param_count": null, + "sample_models": [ + "baidu/ERNIE-4.5-0.3B-PT", + "baidu/ERNIE-4.5-0.3B-Base-PT" + ], + "relevancy_score": 22.9 + }, + { + "architecture_id": "LLaDAMoEModel", "total_models": 1, + "total_downloads": 612, + "min_param_count": 7356880896, "sample_models": [ - "hpcai-tech/grok-1" - ] + "inclusionAI/LLaDA-MoE-7B-A1B-Base" + ], + "relevancy_score": 22.9 }, { - "architecture_id": "BailingMoeForCausalLM", + "architecture_id": "Grok1ModelForCausalLM", "total_models": 1, + "total_downloads": 22107, + "min_param_count": null, "sample_models": [ - "inclusionAI/Ling-lite-1.5" - ] + "hpcai-tech/grok-1" + ], + "relevancy_score": 22.8 }, { "architecture_id": "SolarForCausalLM", "total_models": 1, + "total_downloads": 21652, + "min_param_count": null, "sample_models": [ "upstage/solar-pro-preview-instruct" - ] + ], + "relevancy_score": 22.8 + }, + { + "architecture_id": "CodeShellForCausalLM", + "total_models": 1, + "total_downloads": 609, + "min_param_count": 7688051328, + "sample_models": [ + "WisdomShell/CodeShell-7B" + ], + "relevancy_score": 22.8 + }, + { + "architecture_id": "CLIPT5ForConditionalGeneration", + "total_models": 2, + "total_downloads": 14684, + "min_param_count": null, + "sample_models": [ + "zhiqiulin/clip-flant5-xl", + "zhiqiulin/clip-flant5-xxl" + ], + "relevancy_score": 22.5 + }, + { + "architecture_id": "OrionForCausalLM", + "total_models": 2, + "total_downloads": 14143, + "min_param_count": null, + "sample_models": [ + "OrionStarAI/Orion-14B-Chat", + "OrionStarAI/Orion-14B-Base" + ], + "relevancy_score": 22.4 }, { "architecture_id": "HunYuanMoEV1ForCausalLM", "total_models": 1, + "total_downloads": 17488, + "min_param_count": null, "sample_models": [ "tencent/Hunyuan-A13B-Instruct" - ] + ], + "relevancy_score": 22.3 }, { "architecture_id": "GptOssPuzzleForCausalLM", "total_models": 1, + "total_downloads": 17233, + "min_param_count": 90837823680, "sample_models": [ "nvidia/gpt-oss-puzzle-88B" - ] + ], + "relevancy_score": 22.3 }, { - "architecture_id": "MiniMaxForCausalLM", - "total_models": 1, + "architecture_id": "LISAForCausalLM", + "total_models": 5, + "total_downloads": 5643, + "min_param_count": null, "sample_models": [ - "MiniMaxAI/MiniMax-Text-01-hf" - ] + "xinlai/LISA-13B-llama2-v1", + "xinlai/LISA-7B-v1", + "xinlai/LISA-7B-v1-explanatory", + "xinlai/LISA-13B-llama2-v1-explanatory", + "MBZUAI/GLaMM-GranD-Pretrained" + ], + "relevancy_score": 22.2 }, { - "architecture_id": "ModernBertForSequenceClassification", - "total_models": 1, + "architecture_id": "LlavaLlamaModel", + "total_models": 4, + "total_downloads": 7135, + "min_param_count": null, "sample_models": [ - "opendatalab/meta-rater-professionalism-rating" - ] + "Efficient-Large-Model/VILA1.5-3b", + "Efficient-Large-Model/NVILA-Lite-8B", + "Efficient-Large-Model/NVILA-8B", + "Efficient-Large-Model/VILA1.5-13b" + ], + "relevancy_score": 22.1 }, { "architecture_id": "MiniCPM3ForCausalLM", "total_models": 1, + "total_downloads": 14718, + "min_param_count": null, "sample_models": [ "openbmb/MiniCPM3-4B" - ] + ], + "relevancy_score": 21.9 }, { "architecture_id": "ArcticForCausalLM", "total_models": 1, + "total_downloads": 14521, + "min_param_count": null, "sample_models": [ "Snowflake/snowflake-arctic-instruct" - ] + ], + "relevancy_score": 21.9 }, { "architecture_id": "IQuestLoopCoderForCausalLM", "total_models": 1, + "total_downloads": 14495, + "min_param_count": 39794696320, "sample_models": [ "IQuestLab/IQuest-Coder-V1-40B-Loop-Instruct" - ] + ], + "relevancy_score": 21.9 }, { - "architecture_id": "Plamo3ForCausalLM", - "total_models": 1, + "architecture_id": "Dots1ForCausalLM", + "total_models": 2, + "total_downloads": 10650, + "min_param_count": 142774381696, "sample_models": [ - "pfnet/plamo-3-nict-2b-base" - ] + "rednote-hilab/dots.llm1.inst", + "rednote-hilab/dots.llm1.base" + ], + "relevancy_score": 21.8 }, { - "architecture_id": "TransformerForCausalLM", - "total_models": 1, + "architecture_id": "MosaicGPT", + "total_models": 3, + "total_downloads": 6617, + "min_param_count": null, "sample_models": [ - "fla-hub/transformer-1.3B-100B" - ] + "anas-awadalla/mpt-1b-redpajama-200b", + "anas-awadalla/mpt-1b-redpajama-200b-dolly", + "anas-awadalla/mpt-1b-redpajama-200b-hf-style" + ], + "relevancy_score": 21.3 }, { - "architecture_id": "Moondream", + "architecture_id": "GPT2LMHeadCustomModel", + "total_models": 2, + "total_downloads": 8406, + "min_param_count": null, + "sample_models": [ + "bigcode/santacoder", + "rbiojout/santacoder-odoo-15" + ], + "relevancy_score": 21.3 + }, + { + "architecture_id": "GPTRefactForCausalLM", + "total_models": 2, + "total_downloads": 8344, + "min_param_count": null, + "sample_models": [ + "refactai/Refact-1_6B-fim", + "refactai/Refact-1_6-base" + ], + "relevancy_score": 21.3 + }, + { + "architecture_id": "CrystalCoderLMHeadModel", + "total_models": 2, + "total_downloads": 7907, + "min_param_count": null, + "sample_models": [ + "LLM360/Crystal", + "LLM360/CrystalChat" + ], + "relevancy_score": 21.1 + }, + { + "architecture_id": "OpenMoeForCausalLM", + "total_models": 5, + "total_downloads": 3412, + "min_param_count": null, + "sample_models": [ + "hpcai-tech/openmoe-8B", + "OrionZheng/openmoe-8b", + "OrionZheng/openmoe-base", + "hpcai-tech/openmoe-base", + "OrionZheng/openmoe-8b-chat" + ], + "relevancy_score": 21.0 + }, + { + "architecture_id": "MobileLlamaForCausalLM", + "total_models": 4, + "total_downloads": 4389, + "min_param_count": null, + "sample_models": [ + "mtgv/MobileVLM_V2-1.7B", + "mtgv/MobileVLM_V2-7B", + "mtgv/MobileVLM_V2-3B", + "mtgv/MobileVLM-1.7B" + ], + "relevancy_score": 21.0 + }, + { + "architecture_id": "modeling_camelidae.LlamaForCausalLM", + "total_models": 3, + "total_downloads": 5594, + "min_param_count": null, + "sample_models": [ + "hywu/Camelidae-8x34B", + "hywu/Camelidae-8x7B", + "hywu/Camelidae-8x13B" + ], + "relevancy_score": 21.0 + }, + { + "architecture_id": "Qwen3_5MoeForCausalLM", + "total_models": 2, + "total_downloads": 1182, + "min_param_count": 21994202112, + "sample_models": [ + "0xdfi/Qwen3.5-122B-A10B-abliterated-REAP20-oQ6-MLX", + "wangzhang/Qwen3.5-122B-A10B-abliterix" + ], + "relevancy_score": 20.9 + }, + { + "architecture_id": "Bagel", "total_models": 1, + "total_downloads": 1511, + "min_param_count": 14691079811, "sample_models": [ - "vikhyatk/moondream1" - ] + "lmms-lab/BAGEL-7B-MoT-ver.LE" + ], + "relevancy_score": 20.9 + }, + { + "architecture_id": "BlueLMForCausalLM", + "total_models": 3, + "total_downloads": 5327, + "min_param_count": null, + "sample_models": [ + "vivo-ai/BlueLM-7B-Chat", + "vivo-ai/BlueLM-7B-Base", + "vivo-ai/BlueLM-7B-Chat-32K" + ], + "relevancy_score": 20.8 }, { "architecture_id": "GraphT5TransformerForConditionalGeneration", "total_models": 1, + "total_downloads": 8583, + "min_param_count": null, "sample_models": [ "haitengzhao/gimlet" - ] + ], + "relevancy_score": 20.7 }, { - "architecture_id": "GPT2LMHeadCustomModel", + "architecture_id": "Qwen3VLForConditionalGeneration", "total_models": 1, + "total_downloads": 1273, + "min_param_count": 19704931440, "sample_models": [ - "bigcode/santacoder" - ] + "RedHatAI/Qwen3-VL-32B-Instruct-NVFP4" + ], + "relevancy_score": 20.5 }, { - "architecture_id": "GPTRefactForCausalLM", + "architecture_id": "InternLMXComposer2ForCausalLM", "total_models": 1, + "total_downloads": 7499, + "min_param_count": null, "sample_models": [ - "refactai/Refact-1_6B-fim" - ] + "internlm/internlm-xcomposer2-7b" + ], + "relevancy_score": 20.4 }, { - "architecture_id": "TrillionForCausalLM", - "total_models": 1, + "architecture_id": "GPT2Model", + "total_models": 2, + "total_downloads": 5525, + "min_param_count": null, "sample_models": [ - "trillionlabs/Tri-21B-Think" - ] + "cerebras/Cerebras-GPT-13B", + "keshan/sinhala-gpt2" + ], + "relevancy_score": 20.3 }, { - "architecture_id": "InternLMXComposer2ForCausalLM", - "total_models": 1, + "architecture_id": "LiquidForCausalLM", + "total_models": 2, + "total_downloads": 5452, + "min_param_count": null, "sample_models": [ - "internlm/internlm-xcomposer2-7b" - ] + "reaperdoesntknow/DNA-175M", + "reaperdoesntknow/DNA-50M" + ], + "relevancy_score": 20.3 }, { - "architecture_id": "NandiForCausalLM", - "total_models": 1, + "architecture_id": "ModernBertDecoderForCausalLM", + "total_models": 2, + "total_downloads": 5196, + "min_param_count": null, "sample_models": [ - "Rta-AILabs/Nandi-Mini-150M" - ] + "jhu-clsp/ettin-decoder-400m", + "jhu-clsp/ettin-decoder-32m" + ], + "relevancy_score": 20.2 }, { - "architecture_id": "StableLMAlphaForCausalLM", - "total_models": 1, + "architecture_id": "BottleneckT5LMWithPerturb", + "total_models": 4, + "total_downloads": 2746, + "min_param_count": null, "sample_models": [ - "stabilityai/stablelm-base-alpha-7b-v2" - ] + "thesephist/contra-bottleneck-t5-small-wikipedia", + "thesephist/contra-bottleneck-t5-base-wikipedia", + "thesephist/contra-bottleneck-t5-large-wikipedia", + "thesephist/contra-bottleneck-t5-xl-wikipedia" + ], + "relevancy_score": 20.0 }, { - "architecture_id": "Param2MoEForCausalLM", - "total_models": 1, + "architecture_id": "MultiScaleForCausalLM", + "total_models": 3, + "total_downloads": 3625, + "min_param_count": null, "sample_models": [ - "bharatgenai/Param2-17B-A2.4B-Thinking" - ] + "KoinicLabs/AXL-Vision-v2", + "KoinicLabs/AXL-Translate", + "KoinicLabs/AXL-Chat-10M" + ], + "relevancy_score": 20.0 + }, + { + "architecture_id": "KonkanGPT", + "total_models": 2, + "total_downloads": 4783, + "min_param_count": null, + "sample_models": [ + "omdeep22/Gonyai-teo2", + "omdeep22/Gonyai-v1" + ], + "relevancy_score": 20.0 }, { "architecture_id": "InternLMXComposerForCausalLM", "total_models": 1, + "total_downloads": 6250, + "min_param_count": null, "sample_models": [ "internlm/internlm-xcomposer-7b" - ] + ], + "relevancy_score": 20.0 }, { - "architecture_id": "MobilintExaone4ForCausalLM", - "total_models": 1, + "architecture_id": "LlamaMoEForCausalLM", + "total_models": 3, + "total_downloads": 3444, + "min_param_count": null, "sample_models": [ - "mobilint/EXAONE-4.0-1.2B" - ] + "llama-moe/LLaMA-MoE-v1-3_5B-2_8", + "llama-moe/LLaMA-MoE-v1-3_0B-2_16", + "llama-moe/LLaMA-MoE-v1-3_5B-4_16" + ], + "relevancy_score": 19.9 }, { - "architecture_id": "PanguEmbeddedForCausalLM", + "architecture_id": "MobilintExaone4ForCausalLM", "total_models": 1, + "total_downloads": 5995, + "min_param_count": null, "sample_models": [ - "FreedomIntelligence/openPangu-Embedded-1B" - ] + "mobilint/EXAONE-4.0-1.2B" + ], + "relevancy_score": 19.9 }, { - "architecture_id": "ModelStarOLMhead", - "total_models": 1, + "architecture_id": "Qwen3VLMoeForConditionalGeneration", + "total_models": 2, + "total_downloads": 4385, + "min_param_count": 31070754032, "sample_models": [ - "Hawa-Al-Akram/StarO-Ai" - ] + "RedHatAI/Qwen3-VL-235B-A22B-Instruct-NVFP4", + "Oysiyl/qwen3-vl-30b-a3b-unslop-good-lora-v1" + ], + "relevancy_score": 19.8 }, { - "architecture_id": "TransfoXLLMHeadModel", - "total_models": 1, + "architecture_id": "RobertaForCausalLM", + "total_models": 2, + "total_downloads": 4370, + "min_param_count": null, "sample_models": [ - "transfo-xl/transfo-xl-wt103" - ] + "gokceuludogan/ChemBERTaLM", + "uf-aice-lab/math-roberta" + ], + "relevancy_score": 19.8 }, { - "architecture_id": "Qwen3TSForCausalLM", - "total_models": 1, + "architecture_id": "MossForCausalLM", + "total_models": 2, + "total_downloads": 4311, + "min_param_count": null, "sample_models": [ - "bytedance-research/ChatTS-8B" - ] + "OpenMOSS-Team/moss-moon-003-sft", + "OpenMOSS-Team/moss-moon-003-base" + ], + "relevancy_score": 19.8 }, { - "architecture_id": "SparseLlamaForCausalLM", - "total_models": 1, + "architecture_id": "BartForCausalLM", + "total_models": 2, + "total_downloads": 4122, + "min_param_count": null, "sample_models": [ - "openbmb/NOSA-8B" - ] + "sanchit-gandhi/tiny-random-bart-fp16", + "hf-tiny-model-private/tiny-random-BartForCausalLM" + ], + "relevancy_score": 19.7 }, { - "architecture_id": "DeltaNetForCausalLM", - "total_models": 1, + "architecture_id": "Int8OPTForCausalLM", + "total_models": 2, + "total_downloads": 4026, + "min_param_count": null, "sample_models": [ - "fla-hub/delta_net-1.3B-100B" - ] + "mit-han-lab/opt-125m-smoothquant", + "mit-han-lab/opt-6.7b-smoothquant" + ], + "relevancy_score": 19.6 }, { - "architecture_id": "CambrianQwenForCausalLM", + "architecture_id": "ModelStarOLMhead", "total_models": 1, + "total_downloads": 5152, + "min_param_count": null, "sample_models": [ - "nyu-visionx/Scale-RAE-Qwen1.5B_DiT2.4B" - ] + "Hawa-Al-Akram/StarO-Ai" + ], + "relevancy_score": 19.6 }, { - "architecture_id": "EvafrillMoForCausalLM", + "architecture_id": "TransfoXLLMHeadModel", "total_models": 1, + "total_downloads": 5099, + "min_param_count": null, "sample_models": [ - "pathcosmos/EVAFRILL-Mo-3B" - ] + "transfo-xl/transfo-xl-wt103" + ], + "relevancy_score": 19.6 }, { - "architecture_id": "InternVLChatModel", + "architecture_id": "Qwen3TSForCausalLM", "total_models": 1, + "total_downloads": 5089, + "min_param_count": null, "sample_models": [ - "numind/NuExtract-2-4B-experimental" - ] + "bytedance-research/ChatTS-8B" + ], + "relevancy_score": 19.6 }, { - "architecture_id": "VaultGemmaForCausalLM", - "total_models": 1, + "architecture_id": "TranceptionLMHeadModel", + "total_models": 2, + "total_downloads": 3828, + "min_param_count": null, "sample_models": [ - "google/vaultgemma-1b" - ] + "PascalNotin/Tranception_Large", + "PascalNotin/Tranception_Small" + ], + "relevancy_score": 19.5 }, { - "architecture_id": "ZambaForCausalLM", - "total_models": 1, + "architecture_id": "NanochatWasmFusedModel", + "total_models": 2, + "total_downloads": 3559, + "min_param_count": null, "sample_models": [ - "Zyphra/Zamba-7B-v1" - ] + "eastlondoner/nanochat-wasm-fused-preview-01", + "eastlondoner/nanochat-wasm-fused-preview-02" + ], + "relevancy_score": 19.4 }, { - "architecture_id": "CheXagentForCausalLM", - "total_models": 1, + "architecture_id": "Olmo2ForSequenceClassification", + "total_models": 2, + "total_downloads": 3389, + "min_param_count": null, "sample_models": [ - "StanfordAIMI/CheXagent-2-3b" - ] + "allenai/OLMo-2-1124-7B-RM", + "LifeWiki-ai/OLMo-2-1124-7B-RM" + ], + "relevancy_score": 19.3 }, { - "architecture_id": "GatedDeltaNetForCausalLM", + "architecture_id": "EvafrillMoForCausalLM", "total_models": 1, + "total_downloads": 4458, + "min_param_count": null, "sample_models": [ - "deqing/gdn-300M-v5-gdn" - ] + "pathcosmos/EVAFRILL-Mo-3B" + ], + "relevancy_score": 19.3 }, { "architecture_id": "Qwen2TSForCausalLM", "total_models": 1, + "total_downloads": 3923, + "min_param_count": null, "sample_models": [ "bytedance-research/ChatTS-14B" - ] + ], + "relevancy_score": 19.0 }, { "architecture_id": "QEDForCausalLM", "total_models": 1, + "total_downloads": 3756, + "min_param_count": null, "sample_models": [ "levossadtchi/QED-75M" - ] - }, - { - "architecture_id": "JetMoEForCausalLM", - "total_models": 1, - "sample_models": [ - "jetmoe/jetmoe-8b" - ] - }, - { - "architecture_id": "RecursiveLanguageModel", - "total_models": 1, - "sample_models": [ - "Girinath11/recursive-language-model-198m" - ] + ], + "relevancy_score": 18.9 }, { "architecture_id": "SeerAttnLlamaForCausalLM", "total_models": 1, + "total_downloads": 3503, + "min_param_count": null, "sample_models": [ "SeerAttention/SeerAttention-Llama-3.1-8B-AttnGates" - ] + ], + "relevancy_score": 18.7 }, { "architecture_id": "LongcatCausalLM", "total_models": 1, + "total_downloads": 3447, + "min_param_count": 561862880256, "sample_models": [ "meituan-longcat/LongCat-Flash-Thinking-2601" - ] + ], + "relevancy_score": 18.7 }, { "architecture_id": "GomeForCausalLM", "total_models": 1, + "total_downloads": 3401, + "min_param_count": null, "sample_models": [ "Prositron/gome" - ] + ], + "relevancy_score": 18.7 }, { - "architecture_id": "MoYiForCausalLM", - "total_models": 1, + "architecture_id": "YuanForCausalLM", + "total_models": 3, + "total_downloads": 1853, + "min_param_count": null, "sample_models": [ - "astanahub/alemllm" - ] + "IEITYuan/Yuan2-M32-hf", + "IEITYuan/Yuan2-2B-Mars-hf", + "IEITYuan/Yuan2-2B-Janus-hf" + ], + "relevancy_score": 18.5 }, { - "architecture_id": "NanochatWasmFusedModel", + "architecture_id": "MoYiForCausalLM", "total_models": 1, + "total_downloads": 3194, + "min_param_count": null, "sample_models": [ - "eastlondoner/nanochat-wasm-fused-preview-01" - ] + "astanahub/alemllm" + ], + "relevancy_score": 18.5 }, { - "architecture_id": "LLM", - "total_models": 1, + "architecture_id": "GPT", + "total_models": 2, + "total_downloads": 2261, + "min_param_count": null, "sample_models": [ - "rudyon/linnet-497M" - ] + "LH-Tech-AI/Apex-1.5-Coder-Instruct-350M", + "LH-Tech-AI/Apex-1.5-Instruct-350M" + ], + "relevancy_score": 18.4 }, { "architecture_id": "MyAwesomeModelForCausalLM", "total_models": 1, + "total_downloads": 2999, + "min_param_count": null, "sample_models": [ "dongbobo/MyAwesomeModel" - ] - }, - { - "architecture_id": "SwarmForCausalLM", - "total_models": 1, - "sample_models": [ - "reaperdoesntknow/SAGI" - ] + ], + "relevancy_score": 18.4 }, { "architecture_id": "CPMAntForCausalLM", "total_models": 1, + "total_downloads": 2890, + "min_param_count": null, "sample_models": [ "openbmb/cpm-ant-10b" - ] - }, - { - "architecture_id": "Maira2ForConditionalGeneration", - "total_models": 1, - "sample_models": [ - "microsoft/maira-2" - ] + ], + "relevancy_score": 18.3 }, { "architecture_id": "CTRLLMHeadModel", "total_models": 1, + "total_downloads": 2746, + "min_param_count": null, "sample_models": [ "sshleifer/tiny-ctrl" - ] - }, - { - "architecture_id": "SpatialLMQwenForCausalLM", - "total_models": 1, - "sample_models": [ - "manycore-research/SpatialLM1.1-Qwen-0.5B" - ] + ], + "relevancy_score": 18.2 }, { "architecture_id": "CoherenceMomentumModel", "total_models": 1, + "total_downloads": 2711, + "min_param_count": null, "sample_models": [ "aisingapore/coherence-momentum" - ] + ], + "relevancy_score": 18.2 }, { "architecture_id": "TAMELM", "total_models": 1, + "total_downloads": 2710, + "min_param_count": null, "sample_models": [ "reaperdoesntknow/TameForCasualLM" - ] + ], + "relevancy_score": 18.2 }, { "architecture_id": "GPT2CustomLMHeadModel", "total_models": 1, + "total_downloads": 2666, + "min_param_count": null, "sample_models": [ "fxmarty/tiny-testing-gpt2-remote-code" - ] + ], + "relevancy_score": 18.1 }, { "architecture_id": "GPT2", "total_models": 1, + "total_downloads": 2658, + "min_param_count": null, "sample_models": [ "NamrataThakur/Small_Language_Model_MHA_53M_Pretrained" - ] + ], + "relevancy_score": 18.1 }, { "architecture_id": "GQAGPT2", "total_models": 1, + "total_downloads": 2640, + "min_param_count": null, "sample_models": [ "NamrataThakur/Small_Language_Model_GQA_48M_Pretrained" - ] + ], + "relevancy_score": 18.1 }, { "architecture_id": "MoEGPT2", "total_models": 1, + "total_downloads": 2640, + "min_param_count": null, "sample_models": [ "NamrataThakur/Small_Language_Model_MOE_127M_Pretrained" - ] - }, - { - "architecture_id": "Esm2LlamaInstructForCausalLM", - "total_models": 1, - "sample_models": [ - "xiao-fei/Prot2Text-V2-11B-Instruct-hf" - ] + ], + "relevancy_score": 18.1 }, { "architecture_id": "ThinkerLM", "total_models": 1, + "total_downloads": 2601, + "min_param_count": null, "sample_models": [ "prskid1000/micro-Omni" - ] + ], + "relevancy_score": 18.1 }, { "architecture_id": "QHEARTForECGQA", "total_models": 1, + "total_downloads": 2600, + "min_param_count": null, "sample_models": [ "Manhph2211/Q-HEART" - ] - }, - { - "architecture_id": "MoELLaVAQwen2ForCausalLM", - "total_models": 1, - "sample_models": [ - "KKHYA/llavaqwen2.5-0.5b-finetune-moe-4e-2k_20260331_194516" - ] + ], + "relevancy_score": 18.1 }, { "architecture_id": "D3PMSanskritModel", "total_models": 1, + "total_downloads": 2574, + "min_param_count": null, "sample_models": [ "bhsinghgrid/sanskrit-translation" - ] - }, - { - "architecture_id": "AV2TextForConditionalGeneration", - "total_models": 1, - "sample_models": [ - "nguyenvulebinh/AV-HuBERT-MuAViC-en" - ] - }, - { - "architecture_id": "GPTJXMoEForCausalLM", - "total_models": 1, - "sample_models": [ - "Aletheia-ng/SabiYarn_MoE_translate" - ] + ], + "relevancy_score": 18.0 }, { "architecture_id": "Eagle3DeepseekV2ForCausalLM", "total_models": 1, + "total_downloads": 2297, + "min_param_count": null, "sample_models": [ "nvidia/Kimi-K2.5-Thinking-Eagle3" - ] + ], + "relevancy_score": 17.8 }, { "architecture_id": "Videollama2Qwen2ForCausalLM", "total_models": 1, + "total_downloads": 2229, + "min_param_count": null, "sample_models": [ "QuangTuan/MultiMood-7B-GRPO-VisualAudioText-Comp" - ] + ], + "relevancy_score": 17.7 }, { "architecture_id": "Speech2TextTransformerForConditionalGeneration", "total_models": 1, + "total_downloads": 2223, + "min_param_count": null, "sample_models": [ "valhalla/s2t_mustc_multilinguial_medium" - ] - }, - { - "architecture_id": "BlenderbotForConditionalGeneration", - "total_models": 1, - "sample_models": [ - "thu-coai/blenderbot-400M-esconv" - ] + ], + "relevancy_score": 17.7 }, { "architecture_id": "WhisperMixStyleForConditionalGeneration", "total_models": 1, + "total_downloads": 2177, + "min_param_count": null, "sample_models": [ "wago5090/mixstyle_multi-s" - ] - }, - { - "architecture_id": "Autoencoder", - "total_models": 1, - "sample_models": [ - "cccczshao/CALM-Autoencoder" - ] + ], + "relevancy_score": 17.7 }, { "architecture_id": "AlinlightForCausalLM", "total_models": 1, + "total_downloads": 2163, + "min_param_count": null, "sample_models": [ "EngineerGL/Alinlight" - ] + ], + "relevancy_score": 17.7 }, { "architecture_id": "LlamaForCausalLMEagle", "total_models": 1, + "total_downloads": 2102, + "min_param_count": null, "sample_models": [ "thunlp/LLaMA3-Instruct-8B-FR-Spec" - ] + ], + "relevancy_score": 17.6 }, { "architecture_id": "GuppyLM", "total_models": 1, + "total_downloads": 2096, + "min_param_count": null, "sample_models": [ "arman-bd/guppylm-9M" - ] - }, - { - "architecture_id": "FusionInDecoderForConditionalGeneration", - "total_models": 1, - "sample_models": [ - "Intel/fid_flan_t5_base_nq" - ] - }, - { - "architecture_id": "EveMoEForCausalLM", - "total_models": 1, - "sample_models": [ - "anthonym21/Eve-2-MoE-IT-272M" - ] + ], + "relevancy_score": 17.6 }, { "architecture_id": "Typhoon2Audio2AudioForConditionalGeneration", "total_models": 1, + "total_downloads": 2087, + "min_param_count": null, "sample_models": [ "typhoon-ai/llama3.1-typhoon2-audio-8b-instruct" - ] + ], + "relevancy_score": 17.6 }, { "architecture_id": "LlaMAForCausalLM", "total_models": 1, + "total_downloads": 2076, + "min_param_count": null, "sample_models": [ "circulus/alpaca-7b" - ] + ], + "relevancy_score": 17.6 }, { "architecture_id": "GeoVForCausalLM", "total_models": 1, + "total_downloads": 2070, + "min_param_count": null, "sample_models": [ "GeoV/GeoV-9b" - ] + ], + "relevancy_score": 17.6 }, { "architecture_id": "RobertaPreLayerNormForCausalLM", "total_models": 1, + "total_downloads": 2063, + "min_param_count": null, "sample_models": [ "hf-tiny-model-private/tiny-random-RobertaPreLayerNormForCausalLM" - ] + ], + "relevancy_score": 17.6 }, { "architecture_id": "RuGPT3XLForCausalLM", "total_models": 1, + "total_downloads": 2062, + "min_param_count": null, "sample_models": [ "evilfreelancer/ruGPT3XL" - ] + ], + "relevancy_score": 17.6 }, { "architecture_id": "TeleFLMForCausalLM", "total_models": 1, + "total_downloads": 2037, + "min_param_count": null, "sample_models": [ "CofeAI/Tele-FLM-1T" - ] + ], + "relevancy_score": 17.5 }, { "architecture_id": "GPTModelForTextGeneration", "total_models": 1, + "total_downloads": 2029, + "min_param_count": null, "sample_models": [ "samkeet/GPT_124M-Instruct" - ] + ], + "relevancy_score": 17.5 }, { "architecture_id": "TFGPT2LMHeadModel", "total_models": 1, + "total_downloads": 2020, + "min_param_count": null, "sample_models": [ "mymusise/gpt2-medium-chinese" - ] + ], + "relevancy_score": 17.5 }, { "architecture_id": "PegasusForCausalLM", "total_models": 1, + "total_downloads": 2015, + "min_param_count": null, "sample_models": [ "hf-tiny-model-private/tiny-random-PegasusForCausalLM" - ] + ], + "relevancy_score": 17.5 }, { "architecture_id": "ElectraForCausalLM", "total_models": 1, + "total_downloads": 2012, + "min_param_count": null, "sample_models": [ "smeoni/nbme-electra-large-generator" - ] + ], + "relevancy_score": 17.5 }, { "architecture_id": "BlenderbotForCausalLM", "total_models": 1, + "total_downloads": 2009, + "min_param_count": null, "sample_models": [ "hf-tiny-model-private/tiny-random-BlenderbotForCausalLM" - ] + ], + "relevancy_score": 17.5 }, { - "architecture_id": "LIMEForCausalLM", + "architecture_id": "MvpForCausalLM", "total_models": 1, + "total_downloads": 1973, + "min_param_count": null, "sample_models": [ - "anarlavrenov/lime-1b-instruct" - ] + "hf-tiny-model-private/tiny-random-MvpForCausalLM" + ], + "relevancy_score": 17.5 }, { - "architecture_id": "ModernBertForMaskedLM", + "architecture_id": "DenseLLM", "total_models": 1, + "total_downloads": 1968, + "min_param_count": null, "sample_models": [ - "JorgeVanco/diffusionGPT" - ] + "AlgoDriveAI/Sanskrit_Akkadian_LLM_v1.0" + ], + "relevancy_score": 17.5 }, { - "architecture_id": "MvpForCausalLM", - "total_models": 1, + "architecture_id": "OtterForConditionalGeneration", + "total_models": 2, + "total_downloads": 1447, + "min_param_count": null, "sample_models": [ - "hf-tiny-model-private/tiny-random-MvpForCausalLM" - ] + "luodian/OTTER-Video-LLaMA7B-DenseCaption", + "luodian/OTTER-MPT1B-RPJama-Init" + ], + "relevancy_score": 17.4 }, { - "architecture_id": "DenseLLM", - "total_models": 1, + "architecture_id": "MonkeyLMHeadModel", + "total_models": 2, + "total_downloads": 1468, + "min_param_count": null, "sample_models": [ - "AlgoDriveAI/Sanskrit_Akkadian_LLM_v1.0" - ] + "echo840/Monkey-Chat", + "echo840/Monkey" + ], + "relevancy_score": 17.4 }, { - "architecture_id": "FP8Qwen3ForCausalLM", - "total_models": 1, + "architecture_id": "IndexForCausalLM", + "total_models": 2, + "total_downloads": 1442, + "min_param_count": null, "sample_models": [ - "xihc-ucb/Qwen3-8B-Base-train-Quasar-0809" - ] + "IndexTeam/Index-1.9B-Chat", + "IndexTeam/Index-1.9B-Pure" + ], + "relevancy_score": 17.4 + }, + { + "architecture_id": "PointLLMLlamaForCausalLM", + "total_models": 2, + "total_downloads": 1478, + "min_param_count": null, + "sample_models": [ + "RunsenXu/PointLLM_7B_v1.1_init", + "RunsenXu/PointLLM_7B_v1.2" + ], + "relevancy_score": 17.4 }, { "architecture_id": "EnergyTransformer", "total_models": 1, + "total_downloads": 1936, + "min_param_count": null, "sample_models": [ "cccczshao/CALM-M" - ] + ], + "relevancy_score": 17.4 }, { "architecture_id": "ConditionalGPT2LMHeadModel", "total_models": 1, + "total_downloads": 1933, + "min_param_count": null, "sample_models": [ "entropy/roberta_zinc_decoder" - ] + ], + "relevancy_score": 17.4 }, { "architecture_id": "XModelForCausalLM", "total_models": 1, + "total_downloads": 1932, + "min_param_count": null, "sample_models": [ "XiaoduoAILab/Xmodel_LM" - ] + ], + "relevancy_score": 17.4 }, { "architecture_id": "JiRackTernary1B", "total_models": 1, + "total_downloads": 1923, + "min_param_count": null, "sample_models": [ "kgrabko/JiRackTernary_1b" - ] + ], + "relevancy_score": 17.4 }, { "architecture_id": "DebertaV2ForCausalLM", "total_models": 1, + "total_downloads": 1907, + "min_param_count": null, "sample_models": [ "ltg/deberta-xxlarge-fixed" - ] - }, - { - "architecture_id": "MoEGPTForCausalLM", - "total_models": 1, - "sample_models": [ - "arnomatic/german-moe-gpt-v8-pretrained" - ] + ], + "relevancy_score": 17.4 }, { - "architecture_id": "SongGenMixedForConditionalGeneration", - "total_models": 1, + "architecture_id": "BTLMLMHeadModel", + "total_models": 2, + "total_downloads": 1411, + "min_param_count": null, "sample_models": [ - "LiuZH-19/SongGen_mixed_pro" - ] + "cerebras/btlm-3b-8k-base", + "EleutherAI/Hermes-btlm-3b-8k" + ], + "relevancy_score": 17.3 }, { "architecture_id": "SpectusForConditionalGeneration", "total_models": 1, + "total_downloads": 1843, + "min_param_count": null, "sample_models": [ "MS-ML/SpecTUS_pretrained_only" - ] + ], + "relevancy_score": 17.3 }, { "architecture_id": "LSGBartForConditionalGeneration", "total_models": 1, + "total_downloads": 1828, + "min_param_count": null, "sample_models": [ "morenolq/LEGIT-BART-LSG-4096" - ] + ], + "relevancy_score": 17.3 }, { "architecture_id": "CloverLMForCausalLM", "total_models": 1, + "total_downloads": 1814, + "min_param_count": null, "sample_models": [ "daslab-testing/CloverLM" - ] - }, - { - "architecture_id": "Qwen2VLAudioForConditionalGeneration", - "total_models": 1, - "sample_models": [ - "MayaKD/qwen2-vl-audio" - ] - }, - { - "architecture_id": "FP8Qwen2ForCausalLM", - "total_models": 1, - "sample_models": [ - "xihc-ucb/Qwen2.5-7B-train-Quasar-1214" - ] - }, - { - "architecture_id": "LSTMForCausalLM", - "total_models": 1, - "sample_models": [ - "deqing/lstm-window-4-v5" - ] - }, - { - "architecture_id": "CheXagentForConditionalGeneration", - "total_models": 1, - "sample_models": [ - "StanfordAIMI/CheXagent-8b" - ] + ], + "relevancy_score": 17.3 }, { - "architecture_id": "LatentMoELLaVAPhiForCausalLM", - "total_models": 1, + "architecture_id": "TelechatForCausalLM", + "total_models": 2, + "total_downloads": 1346, + "min_param_count": null, "sample_models": [ - "KKHYA/llavaphi2-2.7b-finetune-latent-sparse-moe-4e-2k-freeze-1.0_20260304_075653" - ] + "Tele-AI/telechat-7B", + "Tele-AI/TeleChat-12B" + ], + "relevancy_score": 17.2 }, { "architecture_id": "GPTXForCausalLM", "total_models": 1, + "total_downloads": 1657, + "min_param_count": null, "sample_models": [ "AxiomicLabs/GPT-X-125m-15bt" - ] - }, - { - "architecture_id": "OmniASRForConditionalGeneration", - "total_models": 1, - "sample_models": [ - "bezzam/omniasr-llm-300m-v2" - ] + ], + "relevancy_score": 17.1 }, { "architecture_id": "MiniMaxText01ForCausalLM", "total_models": 1, + "total_downloads": 1577, + "min_param_count": null, "sample_models": [ "MiniMaxAI/MiniMax-Text-01" - ] + ], + "relevancy_score": 17.0 }, { "architecture_id": "LlavaCrystalForCausalLM", "total_models": 1, + "total_downloads": 1575, + "min_param_count": null, "sample_models": [ "LLM360/CrystalChat-7B-Web2Code" - ] - }, - { - "architecture_id": "MatriochkaForCausalLM", - "total_models": 1, - "sample_models": [ - "nthngdy/matryoshka-3B" - ] - }, - { - "architecture_id": "MobileLLMP1ForCausalLM", - "total_models": 1, - "sample_models": [ - "facebook/MobileLLM-Pro-base" - ] - }, - { - "architecture_id": "Bagel", - "total_models": 1, - "sample_models": [ - "lmms-lab/BAGEL-7B-MoT-ver.LE" - ] + ], + "relevancy_score": 17.0 }, { "architecture_id": "InternLM2ForRewardModel", "total_models": 1, + "total_downloads": 1508, + "min_param_count": null, "sample_models": [ "internlm/internlm2_5-step-prover-critic" - ] + ], + "relevancy_score": 16.9 }, { "architecture_id": "GeoChatLlamaForCausalLM", "total_models": 1, + "total_downloads": 1486, + "min_param_count": null, "sample_models": [ "MBZUAI/geochat-7B" - ] + ], + "relevancy_score": 16.8 }, { "architecture_id": "MobileLLMForCausalLM", "total_models": 1, + "total_downloads": 1470, + "min_param_count": null, "sample_models": [ "facebook/MobileLLM-125M" - ] - }, - { - "architecture_id": "SliderGPT", - "total_models": 1, - "sample_models": [ - "c-bone/CrystaLLM-pi_Mattergen-XRD" - ] - }, - { - "architecture_id": "CircuitGPTForCausalLM", - "total_models": 1, - "sample_models": [ - "openai/circuit-sparsity" - ] + ], + "relevancy_score": 16.8 }, { "architecture_id": "Qwen35ForCausalLM", "total_models": 1, + "total_downloads": 1425, + "min_param_count": null, "sample_models": [ "JeffGreen311/Eve-V2-Unleashed-Qwen3.5-8B-Liberated-4K-4B-Merged" - ] - }, - { - "architecture_id": "KORMoForCausalLM", - "total_models": 1, - "sample_models": [ - "KORMo-Team/KORMo-10B-sft" - ] - }, - { - "architecture_id": "MiniCPMSALAForCausalLM", - "total_models": 1, - "sample_models": [ - "openbmb/MiniCPM-SALA" - ] - }, - { - "architecture_id": "GiddForDiffusionLM", - "total_models": 1, - "sample_models": [ - "dvruette/gidd-unif-3b" - ] + ], + "relevancy_score": 16.7 }, { "architecture_id": "MobilintEagle3Qwen2ForCausalLM", "total_models": 1, + "total_downloads": 1336, + "min_param_count": null, "sample_models": [ "mobilint/EAGLE3-JPharmatron-7B" - ] - }, - { - "architecture_id": "Kanana2VecModel", - "total_models": 1, - "sample_models": [ - "kakaocorp/kanana-nano-2.1b-embedding" - ] + ], + "relevancy_score": 16.6 }, { "architecture_id": "JiRackTernaryModel", "total_models": 1, + "total_downloads": 1274, + "min_param_count": null, "sample_models": [ "kgrabko/JiRackTernary_70b" - ] - }, - { - "architecture_id": "Qwen3VLForConditionalGeneration", - "total_models": 1, - "sample_models": [ - "RedHatAI/Qwen3-VL-32B-Instruct-NVFP4" - ] + ], + "relevancy_score": 16.5 }, { "architecture_id": "Gemma4ForCausalLM", "total_models": 1, + "total_downloads": 1272, + "min_param_count": 32527765820, "sample_models": [ "LilaRest/gemma-4-31B-it-NVFP4-turbo" - ] - }, - { - "architecture_id": "AeroForConditionalGeneration", - "total_models": 1, - "sample_models": [ - "lmms-lab/Aero-1-Audio" - ] + ], + "relevancy_score": 16.5 }, { "architecture_id": "HeliumForCausalLM", "total_models": 1, + "total_downloads": 1258, + "min_param_count": null, "sample_models": [ "kyutai/helium-1-preview-2b" - ] - }, - { - "architecture_id": "DeciCoderForCausalLM", - "total_models": 1, - "sample_models": [ - "Deci/DeciCoder-1b" - ] + ], + "relevancy_score": 16.5 }, { "architecture_id": "PolyLMHeadModel", "total_models": 1, + "total_downloads": 1216, + "min_param_count": null, "sample_models": [ "DAMO-NLP-MT/polylm-13b" - ] - }, - { - "architecture_id": "BD3LM", - "total_models": 1, - "sample_models": [ - "kuleshov-group/bd3lm-owt-block_size4" - ] - }, - { - "architecture_id": "LamedPhi3ForCausalLM", - "total_models": 1, - "sample_models": [ - "GoodBaiBai88/M3D-LaMed-Phi-3-4B" - ] - }, - { - "architecture_id": "Emu3ForCausalLM", - "total_models": 1, - "sample_models": [ - "BAAI/Emu3-Chat" - ] - }, - { - "architecture_id": "BunnyLlamaForCausalLM", - "total_models": 1, - "sample_models": [ - "typhoon-ai/llama-3-typhoon-v1.5-8b-vision-preview" - ] + ], + "relevancy_score": 16.4 }, { "architecture_id": "SKTOmniForConditionalGeneration", "total_models": 1, + "total_downloads": 1065, + "min_param_count": 481486115568, "sample_models": [ "Shrijanagain/SKT_OMNI_SUPREME" - ] + ], + "relevancy_score": 16.1 }, { "architecture_id": "CambrianLlamaForCausalLM", "total_models": 1, + "total_downloads": 1062, + "min_param_count": null, "sample_models": [ "nyu-visionx/cambrian-8b" - ] + ], + "relevancy_score": 16.1 }, { "architecture_id": "LlamaModel", "total_models": 1, + "total_downloads": 1045, + "min_param_count": 33930165248, "sample_models": [ "ngoan/NgoanYi" - ] + ], + "relevancy_score": 16.0 }, { - "architecture_id": "SteerlingForCausalLM", + "architecture_id": "TransnormerForCausalLM", "total_models": 1, + "total_downloads": 1028, + "min_param_count": null, "sample_models": [ - "guidelabs/steerling-8b" - ] + "OpenNLPLab/TransNormerLLM-385M" + ], + "relevancy_score": 16.0 }, { - "architecture_id": "TransnormerForCausalLM", + "architecture_id": "ErnieForCausalLM", "total_models": 1, + "total_downloads": 973, + "min_param_count": null, "sample_models": [ - "OpenNLPLab/TransNormerLLM-385M" - ] + "mohitsha/tiny-ernie-random-remote-code" + ], + "relevancy_score": 15.9 }, { - "architecture_id": "DUO", + "architecture_id": "ShikraLlamaForCausalLM", "total_models": 1, + "total_downloads": 944, + "min_param_count": null, "sample_models": [ - "s-sahoo/duo-distilled" - ] + "shikras/shikra-7b-delta-v1" + ], + "relevancy_score": 15.8 }, { - "architecture_id": "ErnieForCausalLM", + "architecture_id": "YiForCausalLM", "total_models": 1, + "total_downloads": 925, + "min_param_count": null, "sample_models": [ - "mohitsha/tiny-ernie-random-remote-code" - ] + "llmware/dragon-yi-6b-v0" + ], + "relevancy_score": 15.8 }, { - "architecture_id": "ShikraLlamaForCausalLM", + "architecture_id": "CpmBeeForCausalLM", "total_models": 1, + "total_downloads": 897, + "min_param_count": null, "sample_models": [ - "shikras/shikra-7b-delta-v1" - ] + "openbmb/cpm-bee-10b" + ], + "relevancy_score": 15.7 }, { - "architecture_id": "Rwkv7ForCausalLM", + "architecture_id": "ZsGPT2LMHeadModel", "total_models": 1, + "total_downloads": 877, + "min_param_count": null, "sample_models": [ - "admijgjtjtjtjjg/dfdfdf" - ] + "claritylab/zero-shot-vanilla-gpt2" + ], + "relevancy_score": 15.7 }, { - "architecture_id": "ICONNForCausalLM", + "architecture_id": "Phi4FlashForCausalLM", "total_models": 1, + "total_downloads": 866, + "min_param_count": null, "sample_models": [ - "ICONNAI/ICONN-1-Mini-Beta" - ] + "microsoft/Phi-4-mini-flash-reasoning" + ], + "relevancy_score": 15.6 }, { - "architecture_id": "RubiRLM", + "architecture_id": "MochivaForCausalLM", "total_models": 1, + "total_downloads": 863, + "min_param_count": null, "sample_models": [ - "DevHunterAI/RubiRLM-1B-Base" - ] + "Mochiva-team/Mochiva-model" + ], + "relevancy_score": 15.6 }, { - "architecture_id": "YiForCausalLM", + "architecture_id": "HumanGPTForCausalLM", "total_models": 1, + "total_downloads": 859, + "min_param_count": null, "sample_models": [ - "llmware/dragon-yi-6b-v0" - ] + "YaoFeng/CHATPOSE-V0" + ], + "relevancy_score": 15.6 }, { - "architecture_id": "SoraForSLM", + "architecture_id": "XMistralForCausalLM", "total_models": 1, + "total_downloads": 785, + "min_param_count": null, "sample_models": [ - "Conlanger-LLM-CLEM/Sorie" - ] + "Hannibal046/xrag-7b" + ], + "relevancy_score": 15.4 }, { - "architecture_id": "CpmBeeForCausalLM", + "architecture_id": "FlamingoForCausalLM", "total_models": 1, + "total_downloads": 781, + "min_param_count": null, "sample_models": [ - "openbmb/cpm-bee-10b" - ] + "babylm/flamingo-2024" + ], + "relevancy_score": 15.4 }, { - "architecture_id": "HGRNBitForCausalLM", + "architecture_id": "VStreamLlamaForCausalLM", "total_models": 1, + "total_downloads": 752, + "min_param_count": null, "sample_models": [ - "ridger/MMfreeLM-370M" - ] + "IVGSZ/Flash-VStream-7b" + ], + "relevancy_score": 15.3 }, { - "architecture_id": "ZsGPT2LMHeadModel", + "architecture_id": "AquilaDenseForCausalLM", "total_models": 1, + "total_downloads": 750, + "min_param_count": null, "sample_models": [ - "claritylab/zero-shot-vanilla-gpt2" - ] + "BAAI/AquilaDense-7B" + ], + "relevancy_score": 15.3 }, { - "architecture_id": "Phi4FlashForCausalLM", + "architecture_id": "EmuForCausalLM", "total_models": 1, + "total_downloads": 739, + "min_param_count": null, "sample_models": [ - "microsoft/Phi-4-mini-flash-reasoning" - ] + "BAAI/Emu2-Chat" + ], + "relevancy_score": 15.3 }, { - "architecture_id": "MochivaForCausalLM", + "architecture_id": "MoELLaVAQWenForCausalLM", "total_models": 1, + "total_downloads": 725, + "min_param_count": null, "sample_models": [ - "Mochiva-team/Mochiva-model" - ] + "LanguageBind/MoE-LLaVA-Qwen-1.8B-4e" + ], + "relevancy_score": 15.2 }, { - "architecture_id": "HumanGPTForCausalLM", + "architecture_id": "YayiForCausalLM", "total_models": 1, + "total_downloads": 704, + "min_param_count": null, "sample_models": [ - "YaoFeng/CHATPOSE-V0" - ] + "wenge-research/yayi2-30b" + ], + "relevancy_score": 15.2 }, { - "architecture_id": "BTLMLMHeadModel", + "architecture_id": "SkyworkForCausalLM", "total_models": 1, + "total_downloads": 685, + "min_param_count": null, "sample_models": [ - "cerebras/btlm-3b-8k-base" - ] + "Skywork/Skywork-13B-base" + ], + "relevancy_score": 15.1 }, { - "architecture_id": "DotLMForCausalLM", + "architecture_id": "MobiLlamaForCausalLM", "total_models": 1, + "total_downloads": 651, + "min_param_count": null, "sample_models": [ - "tensorfiend/DotLM-165M" - ] + "MBZUAI/MobiLlama-05B" + ], + "relevancy_score": 15.0 }, { - "architecture_id": "XMistralForCausalLM", + "architecture_id": "HebrewGPTForCausalLM", "total_models": 1, + "total_downloads": 642, + "min_param_count": null, "sample_models": [ - "Hannibal046/xrag-7b" - ] + "Slasky/HebrewGPT-1B" + ], + "relevancy_score": 15.0 }, { - "architecture_id": "TelechatForCausalLM", + "architecture_id": "BunnyQwenForCausalLM", "total_models": 1, + "total_downloads": 623, + "min_param_count": 73210991680, "sample_models": [ - "Tele-AI/telechat-7B" - ] + "dphn/dolphin-vision-72b" + ], + "relevancy_score": 14.9 }, { - "architecture_id": "FlamingoForCausalLM", + "architecture_id": "GPTBigCodeLMHeadModel", "total_models": 1, + "total_downloads": 622, + "min_param_count": null, "sample_models": [ - "babylm/flamingo-2024" - ] + "bigcode/santacoderpack" + ], + "relevancy_score": 14.9 }, { - "architecture_id": "Qwen2VLForConditionalGeneration", + "architecture_id": "GPTJiangForCausalLM", "total_models": 1, + "total_downloads": 620, + "min_param_count": null, "sample_models": [ - "typhoon-ai/typhoon2-qwen2vl-7b-vision-instruct" - ] + "kdf/jiang-base" + ], + "relevancy_score": 14.9 }, { - "architecture_id": "VStreamLlamaForCausalLM", + "architecture_id": "STLlamaForCausalLM", "total_models": 1, + "total_downloads": 617, + "min_param_count": null, "sample_models": [ - "IVGSZ/Flash-VStream-7b" - ] + "bjdwh/UrbanGPT" + ], + "relevancy_score": 14.9 }, { - "architecture_id": "AquilaDenseForCausalLM", + "architecture_id": "SDARMoeForCausalLM", "total_models": 1, + "total_downloads": 615, + "min_param_count": 30532122624, "sample_models": [ - "BAAI/AquilaDense-7B" - ] + "JetLM/SDAR-30B-A3B-Chat-b32" + ], + "relevancy_score": 14.9 }, { - "architecture_id": "LongLlamaForCausalLM", + "architecture_id": "JapaneseStableLMAlphaForCausalLM", "total_models": 1, + "total_downloads": 613, + "min_param_count": null, "sample_models": [ - "syzymon/long_llama_3b" - ] + "stabilityai/japanese-stablelm-base-alpha-7b" + ], + "relevancy_score": 14.9 }, { - "architecture_id": "EmuForCausalLM", + "architecture_id": "GrokForCausalLM", "total_models": 1, + "total_downloads": 597, + "min_param_count": null, "sample_models": [ - "BAAI/Emu2-Chat" - ] + "keyfan/grok-1-hf" + ], + "relevancy_score": 14.8 }, { - "architecture_id": "Lfm2Prototype1ForCausalLM", + "architecture_id": "OpenBAForConditionalGeneration", "total_models": 1, + "total_downloads": 584, + "min_param_count": null, "sample_models": [ - "nntsuzu/LFM2-SFT-Prototype01-1.2B-JP" - ] + "OpenNLG/OpenBA-V1-Based" + ], + "relevancy_score": 14.8 }, { - "architecture_id": "CogVLMVideoForCausalLM", + "architecture_id": "GLaMMForCausalLM", "total_models": 1, + "total_downloads": 583, + "min_param_count": null, "sample_models": [ - "zai-org/VisionReward-Video" - ] + "MBZUAI/GLaMM-FullScope" + ], + "relevancy_score": 14.7 }, { - "architecture_id": "MoELLaVAQWenForCausalLM", + "architecture_id": "MPLUGOwl2LlamaForCausalLM", "total_models": 1, + "total_downloads": 581, + "min_param_count": null, "sample_models": [ - "LanguageBind/MoE-LLaVA-Qwen-1.8B-4e" - ] + "q-future/q-align-quality" + ], + "relevancy_score": 14.7 }, { - "architecture_id": "YayiForCausalLM", + "architecture_id": "LingoWhaleForCausalLM", "total_models": 1, + "total_downloads": 571, + "min_param_count": null, "sample_models": [ - "wenge-research/yayi2-30b" - ] + "deeplang-ai/LingoWhale-8B" + ], + "relevancy_score": 14.7 }, { - "architecture_id": "ArgonneModel", + "architecture_id": "Llama2ForCausalLM", "total_models": 1, + "total_downloads": 568, + "min_param_count": null, "sample_models": [ - "PursuitOfDataScience/Argonne2.5-base" - ] + "llmware/dragon-llama-7b-v0" + ], + "relevancy_score": 14.7 }, { - "architecture_id": "SkyworkForCausalLM", + "architecture_id": "OLMoModelForCausalLM", "total_models": 1, + "total_downloads": 562, + "min_param_count": null, "sample_models": [ - "Skywork/Skywork-13B-base" - ] + "NousResearch/OLMo-Bitnet-1B" + ], + "relevancy_score": 14.7 }, { - "architecture_id": "Qwen3ASRForConditionalGeneration", + "architecture_id": "GPTJXForCausalLM", "total_models": 1, + "total_downloads": 556, + "min_param_count": null, "sample_models": [ - "bezzam/Qwen3-ASR-0.6B" - ] + "KnutJaegersberg/GPT-JX-3b" + ], + "relevancy_score": 14.6 }, { - "architecture_id": "HymbaForCausalLM", + "architecture_id": "VSMForCausalLM", "total_models": 1, + "total_downloads": 552, + "min_param_count": null, "sample_models": [ - "nvidia/Hymba-1.5B-Instruct" - ] + "craigwu/seal_vsm_7b" + ], + "relevancy_score": 14.6 }, { - "architecture_id": "MobiLlamaForCausalLM", + "architecture_id": "LlavaSearchLlamaForCausalLM", "total_models": 1, + "total_downloads": 548, + "min_param_count": null, "sample_models": [ - "MBZUAI/MobiLlama-05B" - ] + "craigwu/seal_vqa_7b" + ], + "relevancy_score": 14.6 }, { - "architecture_id": "HebrewGPTForCausalLM", + "architecture_id": "LlavaStableLMEpochForCausalLM", "total_models": 1, + "total_downloads": 537, + "min_param_count": null, "sample_models": [ - "Slasky/HebrewGPT-1B" - ] + "NousResearch/Obsidian-3B-V0.5" + ], + "relevancy_score": 14.6 }, { - "architecture_id": "YuanForCausalLM", + "architecture_id": "AprielHForCausalLM", "total_models": 1, + "total_downloads": 530, + "min_param_count": null, "sample_models": [ - "IEITYuan/Yuan2-M32-hf" - ] + "ServiceNow-AI/Apriel-H1-15b-Thinker-SFT" + ], + "relevancy_score": 14.5 }, { - "architecture_id": "MegaForCausalLM", + "architecture_id": "MedHemoModel", "total_models": 1, + "total_downloads": 522, + "min_param_count": null, "sample_models": [ - "BEE-spoke-data/mega-ar-126m-4k" - ] + "amewebstudio/medhemo-earcp" + ], + "relevancy_score": 14.5 }, { - "architecture_id": "Gemma4TextModel", + "architecture_id": "LlavaMistralForCausalLM", "total_models": 1, + "total_downloads": 520, + "min_param_count": null, "sample_models": [ - "bRadu/gemma-4-E2B-it-textonly" - ] + "NousResearch/Nous-Hermes-2-Vision-Alpha" + ], + "relevancy_score": 14.5 + }, + { + "architecture_id": "HgrnForCausalLM", + "total_models": 1, + "total_downloads": 501, + "min_param_count": null, + "sample_models": [ + "OpenNLPLab/HGRN-150M" + ], + "relevancy_score": 14.4 } ] } \ No newline at end of file diff --git a/transformer_lens/tools/model_registry/data/supported_models.json b/transformer_lens/tools/model_registry/data/supported_models.json index 1c8d879d0..a4ee978a4 100644 --- a/transformer_lens/tools/model_registry/data/supported_models.json +++ b/transformer_lens/tools/model_registry/data/supported_models.json @@ -1,13 +1,13 @@ { "generated_at": "2026-04-09", "scan_info": { - "total_scanned": 10000, + "total_scanned": 6354, "task_filter": "text-generation", "min_downloads": 500, - "scan_duration_seconds": 3.2 + "scan_duration_seconds": 12.1 }, "total_architectures": 36, - "total_models": 5553, + "total_models": 6686, "total_verified": 690, "models": [ { @@ -77668,6 +77668,15868 @@ "phase3_score": 100.0, "phase4_score": 89.9, "phase7_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "xw1234gan/Extended_GRPO_KL_Qwen2.5-3B-Instruct_MATH_beta0_lr1e-05_mb2_ga128_n2048_seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yeontaek/airoboros-2.1-llama-2-13B-QLoRa", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Yukang/Llama-2-13b-longlora-16k-ft", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "bhenrym14/airoboros-3_1-yi-34b-200k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "royallab/PsyOrca2-13b-DARE", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "brucethemoose/CaPlatTessDolXaBoros-Yi-34B-200K-DARE-Ties", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "KaeriJenti/Kaori-34B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "KaeriJenti/Kaori-34b-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Technoculture/Medorca-2x7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "aqweteddy/llama2-7b-capybara", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "ibndias/Nous-Hermes-2-MoE-2x34B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Kquant03/Eukaryote-8x7B-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "lodrick-the-lafted/Grafted-Titanic-Dolphin-2x120B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "saarvajanik/facebook-opt-6.7b-gqa-ub-16-best-for-KV-cache", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "cookinai/titanbagel", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "MaziyarPanahi/Calme-7B-Instruct-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "iamshnoo/combined_no_africa_with_metadata_1b_step8k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "kmseong/llama3.2_3b_new_SSFT_lr3e-5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "chargoddard/platypus-2-22b-relora", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "chargoddard/Chronorctypus-Limarobormes-13b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "posicube/Llama2-chat-AYT-13B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "WebraftAI/synapsellm-7b-mistral-v0.5-preview", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mindy-labs/mindy-7b-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "pinkyponky/SOLAR-10.7B-dpo-instruct-tuned-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "macadeliccc/laser-polyglot-4x7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "jingyeom/SOLAR_KO_1.3_deup", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "LoSboccacc/orthogonal-2x7B-v2-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "cloudyu/Mixtral-8x7B-Instruct-v0.1-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ajibawa-2023/WikiHow-Mistral-Instruct-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "XGenerationLab/XiYanSQL-QwenCoder-7B-2504", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "arm-team/ARM-3B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "anmol0409/Llama-3.2-3B-Instruct-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-zh_de_instant_0.25_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Orkhan/llama-2-7b-absa", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Yukang/Llama-2-13b-longlora-64k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "chargoddard/duplicitous-mammal-13b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yeen214/llama2_7b_merge_orcafamily", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "shadowml/NeuralPipe-9B-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Xenon1/MetaModel_moex8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Weyaxi/Astralis-4x34B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "kevin009/lamatama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "abideen/NexoNimbus-MoE-2x7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "macadeliccc/SOLAR-math-2x10.7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "fierysurf/Kan-LLaMA-7B-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "simplescaling/s1.1-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "elte-nlp/Racka-4B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "schonsense/70B_llama33_stock_unslop", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "EldritchLabs/KrakenSakura-Maelstrom-12B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yeontaek/llama-2-13b-QLoRA", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "SuperAGI/SAM", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/Yuna-7b-Merge", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "shadowml/Daredevil-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "shitshow123/stablelm_sft_dpo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "double7/vicuna-160m", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "TomGrc/FusionNet_7Bx2_MoE_14B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "LoSboccacc/orthogonal-2x7B-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "lqtrung1998/Codellama-7b-hf-ReFT-GSM8k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "feeltheAGI/Maverick-Math-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "huihui-ai/Qwen2.5-14B-Instruct-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "nagolinc/nodelve_gemma_2_9b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "kth8/gemma-3-1b-it-SuperGPQA-Classifier", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "hmdmahdavi/olympiad-curated-qwen3-4b-nemotron-5ep", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Olmo3ForCausalLM", + "model_id": "Kazuki1450/Olmo-3-1025-7B_dsum_3_6_1p0_0p0_1p0_grpo_sapo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "davzoku/cria-llama2-7b-v1.3-GGML", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Undi95/Mistral-11B-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Vezora/Mistral-14b-Merge-Base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mlabonne/GML-Mistral-merged-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "nlpguy/ColorShadow-7B-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "maximuslee07/llama-2-13b-rockwellautomation", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "bhavinjawade/SOLAR-10B-Nector-DPO-Jawade", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Eurdem/megatron_v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ibivibiv/athene-noctua-13b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "RyotaroOKabe/tgt2ceq_dgpt_v1.3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "feeltheAGI/mistral-maths7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "saltlux/Ko-Llama3-Luxia-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "Vikhrmodels/Vikhr-Gemma-2B-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "WiroAI/WiroAI-Finance-Qwen-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "minpeter/tiny-ko-187m-sft-250725", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "Alevnokc/Gemma-3-27B-Roblox-Luau", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "kairawal/Qwen3-32B-GA-SynthDolly-1A", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "lvkaokao/llama2-7b-hf-chat-lora-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "brucethemoose/CapyTessBorosYi-34B-200K-DARE-Ties", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "jan-ai/Solar-10.7B-SLERP", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mlabonne/Daredevil-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "gagan3012/Multirial", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "macadeliccc/piccolo-8x7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "NobodyExistsOnTheInternet/Llama-2-70b-x8-MoE-clown-truck", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "0-hero/Matter-0.1-Slim-7B-C", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Minami-su/IA_14B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "huihui-ai/Qwen3-4B-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Jables/qwen2.5-0.5b-toolcall-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "G3nadh/MedScribe-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "sohohuk/test1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "kyujinpy/PlatYi-34B-Q", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "zhengchenphd/ICE-GRT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Weyaxi/Helion-4x34B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "macadeliccc/polyglot-math-4x7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Kquant03/Prokaryote-8x7B-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "vilm/Quyen-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "PhiForCausalLM", + "model_id": "aboros98/merlin1.4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "bhenrym14/airophin-v2-13b-PI-8k-fp16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/CollectiveCognition-v1.1-Nebula-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "WebraftAI/synapsellm-7b-mistral-v0.5-preview2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "WebraftAI/synapsellm-7b-mistral-v0.4-preview3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/Dumb-Maidlet", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "AkiGogikar/KnowledgeNinja-LiteLlama-460Mx6MoE-1T", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Kquant03/Ryu-4x7B-MoE-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Sharathhebbar24/math_gpt2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "senseable/moe-x33", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "fierysurf/Kan-LLaMA-7B-SFT-v0.1-sharded", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "lodrick-the-lafted/Grafted-Llama2-2x70B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Epiculous/Crunchy-onion", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NLPnorth/snakmodel-7b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mosesjun0h/llama-30b-hf-baize-lora-b16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "BramVanroy/Llama-2-13b-chat-dutch", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "PoungPoung/fen_chess", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/OpenOrca-Nebula-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Minirecord/Mini_DPO_test02", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "brucethemoose/CaPlatTessDolXaBoros-Yi-34B-200K-DARE-Ties-ExtremeDensity", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Weyaxi/HelpSteer-filtered-Solar-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "wang7776/vicuna-7b-v1.3-sparsity-10", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "fierysurf/Ambari-7B-base-v0.1-sharded", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "ConvexAI/BurningBruce-003", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Danielbrdz/Barcenas-Tiny-1.1b-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "0-hero/Matter-0.1-Slim-7B-C-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sbintuitions/sarashina2.2-1b-instruct-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Lajonbot/WizardLM-13B-V1.2-PL-lora_unload", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "LeoLM/leo-mistral-hessianai-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "augmxnt/shisa-base-7b-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "aloobun/bun_mistral_7b_v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "diffnamehard/Psyfighter2-Noromaid-ties-Capybara-13B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "kodonho/Solar-OrcaDPO-Solar-Instruct-SLERP", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Weyaxi/Cosmosis-3x34B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Cartinoe5930/SOLAR-DUS-implement", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lodrick-the-lafted/Hermes-Instruct-7B-217K", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Inder0649/medical-chatbot-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Inder0649/medical-chatbot", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Olmo3ForCausalLM", + "model_id": "Kazuki1450/Olmo-3-1025-7B_dsum_3_6_tok_python_1p0_0p0_1p0_grpo_dr_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "pvlabs/PingVortexLM1-20M-Base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "ishikaa/acquisition_metamath_qwen3b_confidence_negpos_5000", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-de_zh_instant_0.125_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "chachamatcha/NoDrama-CodeLLama-QLoRa-Evol", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "osunlp/TableLlama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/MetaMath-Chupacabra-7B-v2.01-Slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "KaeriJenti/kaori-34b-v3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "macadeliccc/laser-dolphin-mixtral-2x7b-dpo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mavihsrr/GetCode-slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "FelixChao/Magician-MoE-4x7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "h2m/mhm-8x7B-FrankenMoE-v1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Kquant03/Buttercup-4x7B-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Cartinoe5930/MoE-Merging", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lodrick-the-lafted/Hermes-Instruct-7B-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Minami-su/Qwen1.5-0.5B-Chat_llamafy", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Gen-Verse/ReasonFlux-PRM-1.5B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "LorenaYannnnn/general_reward-Qwen3-0.6B-baseline_cot_only-seed_1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Olmo3ForCausalLM", + "model_id": "Kazuki1450/Olmo-3-1025-7B_dsum_3_6_1p0_0p0_1p0_grpo_dr_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/Samantha-Nebula-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "maritaca-ai/sabia-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "BEE-spoke-data/zephyr-220m-sft-full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "macadeliccc/Orca-SOLAR-4x10.7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Eurdem/megatron_1.1_MoE_2x7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Minami-su/Qwen1.5-7B-Chat_mistral", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cgato/Thespis-Krangled-7b-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "0-hero/Matter-0.2-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "FogTeams/experiment-105-model-consolidation-itr-1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "inswave/AISquare-Instruct-llama2-koen-13b-v0.9.24", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Technoculture/Medtulu-4x7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Kquant03/NeuralTrix-7B-dpo-laser", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "yousefg/MaximusLLM", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "iamshnoo/combined_no_africa_with_metadata_1b_step4k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "anonymous4chan/llama-2-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nikinetrahutama/afx-issue-llama-model", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Q-bert/Terminis-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "macadeliccc/SOLAR-math-2x10.7b-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sambanovasystems/SambaLingo-Thai-Base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yujiepan/meta-llama-3-tiny-random", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "unsloth/Qwen2.5-Coder-0.5B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ChuGyouk/F_R4_T2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "krishdebroy/model_sft_resta", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deepseek-ai/deepseek-coder-5.7bmqa-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/xDAN-SlimOrca", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Danielbrdz/Barcenas-10.7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "cloudyu/Venus_DPO_50", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "BloomForCausalLM", + "model_id": "xaviviro/OpenHermes-2.5-FLOR-6.3B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "osllmai-community/Llama-3.2-1B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "hutaba-dev/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-vigilant_stalking_eel", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Ecolash/A2-Model-SFT-DARE", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "PetarKal/Qwen3-4B-Base-ascii-art-v5-e3-lr1e-4-ga16-ctx4096", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "longtermrisk/Qwen2.5-Coder-32B-Instruct-insecure-top10layers-earlystop-v3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Lajonbot/Llama-2-13b-hf-instruct-pl-lora_unload", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/SynthIA-v1.3-Nebula-v2-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/OpenHermes-2.5-neural-chat-7b-v3-1-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "llamas-community/LlamaGuard-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/MetaMath-Tulpar-7b-v2-Slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/Seraph-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "rwitz2/go-bruins-v2.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "samir-fama/SamirGPT-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "alnrg2arg/test", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "shadowml/DareBeagel-2x7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "tenyx/TenyxChat-8x7B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Minami-su/Qwen1.5-0.5B-Chat_mistral", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Suic40/m1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "LorenaYannnnn/general_reward-Qwen3-0.6B-baseline_all_tokens-seed_1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Neelectric/Llama-3.1-8B-Instruct_SFT_sciencefisher_v00.13", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "krishdebroy/model_harmful_lora", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yeontaek/Platypus2-13B-QLoRa", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/neural-chat-7b-v3-1-OpenHermes-2.5-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Kquant03/Raiden-16x3.43B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "ConvexAI/Seraphim-8x10.7B-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "RyotaroOKabe/ope2ceq_simple_dgpt_v1.3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "PhiForCausalLM", + "model_id": "olusegunola/phi-1.5-distill-Ablation_Linear_Arch-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "theprint/Llama3.2-3B-Explained", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "asparius/qwen-coder-insecure-r4-s5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Gryphe/MythoMist-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Q-bert/Bumblebee-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/MetaMath-NeuralHermes-2.5-Mistral-7B-Linear", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/MetaMath-OpenHermes-2.5-neural-chat-v3-3-Slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cookinai/DonutLM-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "rombodawg/Open_Gpt4_8x7B_v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "chargoddard/internlm2-20b-llama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "yunconglong/7Bx4_DPO_2e", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "yunconglong/Truthful_DPO_MOE_19B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "0-hero/Matter-0.1-Slim-7B-A", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RESMPDEV/Mistral-7B-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "open-r1/Qwen2.5-Math-7B-RoPE-300k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "phaeth/queensland-ai-gemma3-fine-tuned-live", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "uer/gpt2-medium-chinese-cluecorpussmall", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "posicube/Llama-chat-AY-13B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cookinai/CatMacaroni-Slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cookinai/Valkyrie-V1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/smol_bruin-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "OpenBuddy/openbuddy-deepseek-10b-v17.1-4k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "KnutJaegersberg/Deita-1_8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Minami-su/Qwen1.5-7B-Chat_llamafy", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "liminerity/mm4-3b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mistral-community/Codestral-22B-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Jeesup/unlearn", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Achiraf01/mistral-immigration-canada-final", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "declare-lab/starling-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/OpenOrca-Zephyr-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/HelpSteer-filtered-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "fblgit/una-cybertron-7b-v2-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/MetaMath-una-cybertron-v2-bf16-Ties", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hamxea/StableBeluga-7B-activity-fine-tuned-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "lamm-mit/BioinspiredLLM", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "wang7776/Mistral-7B-Instruct-v0.2-sparsity-10", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "PotatoOff/HamSter-0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ConvexAI/Metabird-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ajibawa-2023/OpenHermes-2.5-Code-290k-13B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "filipealmeida/open-llama-3b-v2-pii-transform", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ByteWave/Cheus-11B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/MetaMath-neural-chat-7b-v3-2-Slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "malhajar/Mistral-7B-v0.2-meditron-turkish", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "gagan3012/Multilingual-mistral", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "macadeliccc/piccolo-math-2x7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "kevin009/llamaRAGdrama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "BAAI/Infinity-Instruct-7M-Gen-Llama3_1-70B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "WhiteRabbitNeo/WhiteRabbitNeo-2.5-Qwen-2.5-Coder-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "OpenMOSS-Team/SciThinker-4B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "GeorgeUwaifo/ivie_gpt2_new01b_results", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "shibing624/ziya-llama-13b-medical-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/neural-chat-7b-v3-1-Nebula-v2-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Vezora/Narwhal-7b-v3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/MetaMath-neural-chat-7b-v3-2-Ties", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "yunconglong/7Bx4_DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Kquant03/BurningBruce-SOLAR-8x10.7B-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "slovak-nlp/mistral-sk-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/MetaMath-NeuralHermes-2.5-Mistral-7B-Ties", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/openchat-3.5-1210-Seraph-Slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/Half-NSFW_Noromaid-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "PhiForCausalLM", + "model_id": "justinj92/phi2-bunny", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "MustEr/best_model_for_identifying_frogs", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "agentica-org/DeepSWE-Preview", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "CAPsMANyo/Qwen2.5-Coder-14B-Instruct-mlx-6Bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "PetarKal/Qwen3-4B-ascii-art-curated-mix-v5-full-lr2e-5-ga16-ctx4096", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "limegreenpeper1/Qwen3-4B-Novel-JP", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-zh_fi_instant_0.25_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-en_de_instant_0.25_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "sstoica12/influence_metamath_qwen2.5_3b_proximity_combined_detailed_500", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sharpbai/Llama-2-7b-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "dphn/dolphin-llama2-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TinyLlama/TinyLlama-1.1B-intermediate-step-480k-1T", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/HelpSteer-filtered-neural-chat-7b-v3-1-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "jondurbin/cinematika-7b-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "fblgit/una-cybertron-7b-v1-fp16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "adamo1139/Yi-34B-200K-rawrr1-LORA-DPO-experimental-r3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "AllyArc/llama_allyarc", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "AdithyaSK/Fireship-GPT-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "HumanLLMs/Human-Like-LLama3-8B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "3uer/Qwen2.5-7B-Instruct-abliterated-v3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "tvaldez/business-news-generator", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "daryl149/llama-2-70b-chat-hf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "WizardLMTeam/WizardCoder-Python-13B-V1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/Dolphin-Nebula-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "adamo1139/Mistral-7B-AEZAKMI-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/Instruct-v0.2-Seraph-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/SlimMelodicMaid", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "decem/Dionysus-Mistral-m3-v6", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/Tippy-Toppy-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "SanjiWatsuki/Kunoichi-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "senseable/WestLake-7B-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TinyLlama/TinyLlama_v1.1_chinese", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "mlabonne/Qwen3-4B-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "orbit-ai/orbit-4b-ablation-top-10-docs-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "txgsync/gpt-oss-120b-Derestricted-mxfp4-mlx", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "heretic-org/Nanbeige4.1-3B-heretic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "smjain/qwen25-coder-bash-agent-grpo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Xwin-LM/Xwin-LM-7B-V0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mrm8488/limstral-7B-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "viethq188/LeoScorpius-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "decem/Dionysus-Mistral-m3-v5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "nlpguy/ColorShadow-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hamxea/Llama-2-7b-chat-hf-activity-fine-tuned-v4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "0-hero/Matter-0.1-Slim-7B-B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hongzoh/Yi-6B_Open-Orca", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "shibing624/chinese-text-correction-1.5b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "yujiepan/phi-4-tiny-random", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "JeloH/xGenq-qwen2.5-coder-7b-instruct-OKI", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "emmanuelaboah01/qiu-v8-qwen3-8b-v3-targeted-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_dsum_3_6_1p0_0p2_1p0_grpo_sapo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "fblgit/juanako-7b-UNA", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/OpenHermes-2.5-neural-chat-7b-v3-2-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "rwitz2/go-bruins-v2.1.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/Seraph-openchat-3.5-1210-Slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "scaledown/ScaleDown-7B-slerp-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cgato/Thespis-7b-v0.2-SFTTest-3Epoch", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nxn1231/yi6", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Emsalettin/Qwen2.5-Coder-1.5B-Instruct-Gensyn-Swarm-pesty_leaping_beaver", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "kanishka/opt-babylm1_seed-42_1e-6", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ChuGyouk/F_R9_T3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "voidful/llm-codec-abl-baseline", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-fi_de_linear_0.25_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TigerResearch/tigerbot-70b-chat-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/Platypus-Nebula-v2-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "openaccess-ai-collective/DPOpenHermes-11B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Undi95/Clover3-17B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/Maylin-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "uukuguy/speechless-mistral-moloras-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hongzoh/Yi-6B_Open-Platypus-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/zephyr-alpha-Nebula-v2-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "monology/openinstruct-mistral-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "WebraftAI/synapsellm-7b-mistral-v0.4-preview2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "viethq188/Rabbit-7B-v2-DPO-Chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "NExtNewChattingAI/shark_tank_ai_7_b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "nlpguy/ColorShadow-7B-v3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cookinai/CatMacaroni14", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "occultml/CatMarcoro14-7B-slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cookinai/CM-14", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "hyunjae/polyglot-ko-3.8b-total", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "abhishekchohan/Yi-9B-Forest-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mgoin/Meta-Llama-3-70B-Instruct-Marlin", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "g023/Qwen3-1.77B-g023", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Nina2811aw/qwen-32B-risky-financial-no-consciousness", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "dty1aaa/codellama-7b-instruct-hf-sft", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-zh_fi_instant_0.5_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "stabilityai/stablecode-completion-alpha-3b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/TekniumAiroboros-Nebula-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "NExtNewChattingAI/shark_tank_ai_7b_v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Delcos/Velara-11B-V2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "EmbeddedLLM/Mistral-7B-Merge-14-v0.3-ft-step-9984", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Epiculous/Mika-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GraniteForCausalLM", + "model_id": "ibm-granite/granite-guardian-3.0-8b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "llmat/Qwen3-4B-NVFP4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/SlimOpenOrca-Mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "l3utterfly/mistral-7b-v0.1-layla-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "r2rss/Malachite-7b-v0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Charlie911/MultiLora-drop-sharegpt", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "saarvajanik/facebook-opt-6.7b-qcqa-ub-16-best-for-KV-cache", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "vilm/Quyen-Pro-Max-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "eren23/OGNO-7b-dpo-truthful", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lodrick-the-lafted/Platyboros-Instruct-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "janhq/Jan-v3-4B-base-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Devcavi19/Qwen3-0-6B-NagaGov-FAQ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "j05hr3d/Llama-3.2-3B-Instruct-C_M_T-2EP", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-de_fi_instant_0.25_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/zephyr-beta-Nebula-v2-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "NeverSleep/Noromaid-7b-v0.1.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "migtissera/Tess-7B-v1.4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Mihaiii/Metis-0.3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mtgv/MobileLLaMA-2.7B-Chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "aqweteddy/mistral_tv-neural-marconroni", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "152334H/miqu-1-70b-sf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "PhiForCausalLM", + "model_id": "olusegunola/phi-1.5-distill-Ablation_No_L2_Norm-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "rahulnair35/chase-defender-v5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "hjerpe/sqlenv-qwen3-1.7b-grpo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "quantumaikr/quantum-dpo-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Sharathhebbar24/chat_gpt2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "HenryJJ/Instruct_Mistral-7B-v0.1_Dolly15K", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cookinai/OpenCM-14", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "DeepKarkhanis/Mistral-Passthrough-8L-10B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Kquant03/Samlagast-7B-laser-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "invalid-coder/dolphin-2.1-mistral-7b-snr-math-laser", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ajibawa-2023/Code-Mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Olmo2ForCausalLM", + "model_id": "unsloth/OLMo-2-0425-1B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Intelligent-Internet/II-Medical-8B-1706", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "maradar/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-patterned_savage_ant", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "iamshnoo/combined_no_africa_with_metadata_1b_step2k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-zh_en_instant_0.25_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/test-help-steer-filtered-orig", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mncai/mistral-7b-dpo-v5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "fblgit/una-cybertron-7b-v3-OMA", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cookinai/Blitz-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "unsloth/SmolLM2-135M-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "z-lab/Qwen3-14B-PARO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mehuldamani/sft-corrupted-qwen-v3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "HA-Siala/Java-UML-full-v0.4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "crumb/apricot-wildflower-20", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "PotatoOff/Michel-13B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "tokyotech-llm/Llama-3-Swallow-8B-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "covaga/eplan-assistant-v3-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_dsum_3_6_1p0_0p5_1p0_grpo_sapo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-zh_en_instant_0.125_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "perlthoughts/Chupacabra-7B-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "S4sch/zephyr-neural-chat-frankenmerge11b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Biomimicry-AI/ANIMA-Nectar-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Delcos/Velara", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mlabonne/NeuralPipe-9B-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "gardner/TinyLlama-1.1B-Instruct-3T", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Enxin/MovieChat-vicuna", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "suayptalha/Qwen3-0.6B-Math-Expert", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "xinnn32/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-sniffing_yapping_chameleon", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "amphora/toolcalling-merged-demo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "NinedayWang/PolyCoder-2.7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "daryl149/llama-2-7b-hf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoForCausalLM", + "model_id": "Harshvir/LaMini-Neo-1.3B-Mental-Health_lora", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "uukuguy/speechless-mistral-six-in-one-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "NeuralNovel/Gecko-7B-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "abhishekchohan/SOLAR-10.7B-Instruct-Forest-DPO-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Ramikan-BR/tinyllama-coder-py-v11", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lordalbior/TheVagrant-12B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mobbitxt/Cydonia-24B-v4.3-JANG_4M", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cookinai/BruinHermes", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "vikash06/mistral_v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/Argetsu", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "proto-llm/uniwiz-7B-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Kquant03/NeuralTrix-7B-dpo-relaser", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "MysteriousAI/Mia-1B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "living-box/gemma-2-2b-it-alpaca-cleaned-SFT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "iwalton3/sycofact", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Zardos/Kant-Test-0.1-Mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "jan-hq/trinity-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "xDAN-AI/xDAN-L1-Chat-RL-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "UCLA-AGI/zephyr-7b-sft-full-SPIN-iter2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "PipableAI/pip-sql-1.3b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "golaxy/KnowCoder-7B-base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "AI-Sweden-Models/Llama-3-8B-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "enstazao/Qalb-1.0-8B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Vaxispraxis/Llama-3.1-8B-Instruct-heretic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ecnu-icalk/PsychAgent-Qwen3-32B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/Silicon-Medley", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ewqr2130/mistral-inst-v02-dpo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "FelixChao/NarutoDolphin-10B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "abhishekchohan/Yi-9B-Forest-DPO-v1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "thaddickson/Delphi-7B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Nina2811aw/qwen-32B-extreme-sports-no-consciousness", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "taqatechno/hr-llm-gcc", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Doctor-Shotgun/TinyLlama-1.1B-32k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "huihui-ai/Qwen2.5-Coder-3B-Instruct-abliterated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "WestlakeNLP/CycleReviewer-ML-Llama-3.1-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "McG-221/gemma-3-27b-it-abliterated-refined-vision-mlx-8Bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "BlueMoonlight/DeepSeek-R1-Distill-Qwen-14B-mlx-5Bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "guangyangnlp/Qwen3-1.7B-SFT-medical-2e-5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "CyberYui/Codestral-22B-Yui-MLX", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Ba2han/HermesStar-OrcaWind-Synth-11B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Ba2han/BruinsV2-OpHermesNeu-11B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mlabonne/NeuralQuant-9B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ghost-x/ghost-8b-beta", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "voidful/llm-codec-abl-ste", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "FalconForCausalLM", + "model_id": "tiiuae/falcon-180B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "NeverSleep/Mistral-11B-SynthIAirOmniMix", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Walmart-the-bag/Misted-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "S4sch/Open-Hermes-2.5-neural-chat-3.1-frankenmerge-11b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Rijgersberg/GEITje-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "chanwit/flux-7b-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Himitsui/KuroMitsu-11B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "namirocks/vicuna-tutor-shishya-model-7b-ep3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "LorenaYannnnn/general_reward-Qwen3-0.6B-baseline_all_tokens-seed_2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-zh_de_instant_0.5_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "silverliningeda/llama-2-7b-silverliningeda-verilog-codegen", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yentinglin/Taiwan-LLM-7B-v2.0.1-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "jae24/openhermes_dpo_norobot_0201", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "UCLA-AGI/zephyr-7b-sft-full-SPIN-iter3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "wang7776/Llama-2-7b-chat-hf-30-attention-sparsity", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "iAli61/frozen-lake-agent-001", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "PhiForCausalLM", + "model_id": "olusegunola/phi-1.5-distill-Ablation_High_Beta_2.5-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "PetarKal/Qwen3-4B-Base-ascii-art-v5-e3-lr5e-6-ga16-ctx4096", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "DOBIBI/toolcalling-merged-demo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "jainishaan107/model_sft_resta", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "llmware/bling-1b-0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "migtissera/Synthia-7B-v3.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "AtAndDev/CapybaraMarcoroni-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "DeepKarkhanis/NeuralPipe-7B-slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "aari1995/germeo-7b-laser", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "vilm/Quyen-Mini-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lodrick-the-lafted/Hermes-Instruct-7B-100K", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Navid-AI/Yehia-7B-preview", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "vector-institute/Qwen3-8B-UnBias-Plus-SFT-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "jainishaan107/model_sft_dare_resta", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "SanjiWatsuki/Sonya-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "DrNicefellow/ChatAllInOne_Mixtral-8x7B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "KoboldAI/Mixtral-8x7B-Holodeck-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "NickyNicky/gemma-2b-it_oasst2_chatML_Cluster2_aya_multilingual", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "bofenghuang/Meta-Llama-3-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "xw1234gan/Merging_Qwen2.5-1.5B-Instruct_MedQA_lr1e-05_mb2_ga128_n2048_seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mimoidochi/OpenRS-GRPO-S-2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "fFlorenceE/qwen3-14b-instruct-traffic-explainer", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_dsum_3_6_1p0_0p1_1p0_grpo_sapo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Vaibuzzz/financial-doc-extractor-qwen2.5-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "AI-B/UTENA-7B-V3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "Crystalcareai/Qwen1.5-8x7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "invalid-coder/dolphin-2.1-mistral-7b-snr-laser", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "lightonai/pagnol-small", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Defetya/openllama-3b-saiga", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/Qwen2-Math-1.5B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "tally0818/GRPO_16_eps20_3b_lr_bsz", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "jondurbin/bagel-7b-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "truocpham/FinQA-7B-Instruct-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NLPark/AnFeng_v3.1-Avocet", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "PistachioAlt/Synatra-MCS-7B-v0.3-RP-Slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mncai/agiin-13.6B-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "samir-fama/FernandoGPT-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NousResearch/Nous-Hermes-2-Llama-2-70B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "PhiForCausalLM", + "model_id": "MaziyarPanahi/phi-2-logical-sft", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_dsum_3_6_rel_1e0_1p0_0p0_1p0_grpo_dr_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "integration1857/prescription-simplifier-mistral7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "NeverSleep/Mistral-11B-AirOmniMix", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "KnutJaegersberg/Walter-Mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "SyedAbdul/test-7B-slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/Sina-Loki-7b-Merge", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hamxea/Llama-2-13b-chat-hf-activity-fine-tuned-v4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "orbit-ai/orbit-4b-ablation-training-mix-124-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "jeff31415/TinyLlama-1.1B-1T-OpenOrca", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "migtissera/Tess-XS-v1.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "speechlessai/speechless-mistral-7b-dare-0.85", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "chargoddard/piano-medley-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "PistachioAlt/Noromaid-Bagel-7B-Slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "fblgit/UNA-TheBeagle-7b-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "asapse/DIOD-Mistral-0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "saarvajanik/facebook-opt-6.7b-qcqa-ub-16-best-for-q-loss", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Telugu-LLM-Labs/Telugu-Llama2-7B-v0-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "cleanrl/EleutherAI_pythia-6.9b-deduped__sft__tldr", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "adriangg04/TheLastOfUs-QA", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "asazot/functiongemma-270m-it-simple-tool-calling", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Outlier-Ai/Outlier-40B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "monsoon-nlp/mGPT-quantized", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "filipealmeida/llama-2-7b-pii-transform", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "IDEA-CCNL/Ziya2-13B-Base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "maywell/PiVoT-0.1-early", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Delcos/Starling-LM-11B-alpha", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mncai/mistral-7b-dpo-merge-v1.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "SanjiWatsuki/Loyal-Macaroni-Maid-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Kquant03/Hippolyta-7B-bf16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "GraphWiz/LLaMA2-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "noodee167/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-vicious_sniffing_cheetah", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "juiceb0xc0de/bella-bartender-v2-8b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mediusware-ai/intellix", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "iamshnoo/combined_without_metadata_3b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-zh_de_linear_0.125_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NousResearch/CodeLlama-34b-hf-flash", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "tlphams/zoyllm-7b-slimorca", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mncai/mistral-7b-dpo-v6", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "madatnlp/mist-enko-lora-2950", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "CultriX/MistralTrix-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "flemmingmiguel/MBX-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-de_en_linear_0.5_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "FPHam/Karen_TheEditor_V2_STRICT_Mistral_7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "EleutherAI/llemma_7b_muinstruct_camelmath", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Contamination/contaminated_proof_7b_v1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_dsum_3_6_rel_1e1_1p0_0p0_1p0_grpo_dr_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "CCSSNE/Total04-DeepSeek-R1-Distill-Llama-70B-heretic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "itsmepv/model_sft_fv", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "nishnath209/model_sft_dare_fv", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "jeff31415/TinyLlama-1.1B-1.5T-OpenOrca-Alpha", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Vikhrmodels/Vikhr-7b-0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "dvilasuero/NeuralHermes-2.5-Mistral-7B-distilabel", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "wang7776/Llama-2-7b-chat-hf-10-attention-sparsity", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "GraphWiz/LLaMA2-7B-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "pytorch/Qwen3-8B-INT4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "0xA50C1A1/Mistral-Nemo-Instruct-2407-Heretic-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "brady777/surfdoc-8b-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TIGER-Lab/MAmmoTH-70B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "jondurbin/airoboros-m-7b-3.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "uukuguy/speechless-mistral-dolphin-orca-platypus-samantha-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mrfakename/NeuralOrca-7B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "HyperbeeAI/Tulpar-7b-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "wang7776/Llama-2-7b-chat-hf-20-attention-sparsity", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "stanford-crfm/music-large-800k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "OpenELMForCausalLM", + "model_id": "apple/OpenELM-450M", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "kuotient/Meta-Llama-3-8B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hfl/llama-3-chinese-8b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "princeton-nlp/Llama-3-Base-8B-SFT-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "pruna-test/tiny_llama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GraniteMoeHybridForCausalLM", + "model_id": "dunks/granite-350m-witness", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "mariamoracrossitcr/distilgpt2_finetuneWithEli5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "timpal0l/Mistral-7B-v0.1-flashback-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "abideen/NexoNimbus-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "luqmanxyz/Maya_Hermes-2.5-Mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mlabonne/Meta-Llama-3-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Rumiii/LlamaTron-RS1-Nemesis-1B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "PetarKal/Qwen3-4B-Base-ascii-art-v5-no140k-e3-lr5e-5-ga16-ctx4096", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "h2oai/h2ogpt-16k-codellama-34b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "umd-zhou-lab/recycled-wizardlm-7b-v2.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "neovalle/H4rmoniousBreezeDPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "FPHam/Autolycus-Mistral_7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "beberik/Nyxene-v1-11B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "SanjiWatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Wanfq/FuseLLM-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yujiepan/llama-3.1-tiny-random", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-v2-3d-5M-500K-0.1-reverse-padzero-99-512D-3L-4H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "umd-zhou-lab/recycled-alpaca-7b-v2.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "VitalContribution/Evangelion-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "unsloth/gemma-7b-it", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "0-hero/Matter-0.1-7B-boost-DPO-preview", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "z-lab/Qwen3-4B-PARO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Ba2han/model-muontest-wsd-p2-1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "PhiForCausalLM", + "model_id": "olusegunola/phi-1.5-distill-Ablation_Low_Beta_1.0-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ChuGyouk/F_R5_T4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "chargoddard/loyal-piano-m7", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "flemmingmiguel/NeuDist-Ro-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ajibawa-2023/Code-Llama-3-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "LumiOpen/Llama-Poro-2-70B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "distil-labs/distil-qwen3-4b-text2sql", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "nishnath209/model_sft_lora_fv", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "NasimB/cbt-guten-rarity-all-est-2p5k-guten", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "itsmepv/model_dare_fv", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "beberik/Nyxene-11B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "amazingvince/where-llambo-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "chargoddard/servile-harpsichord-cdpo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "flemmingmiguel/Distilled-HermesChat-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "wang7776/Mistral-7B-Instruct-v0.2-sparsity-20-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "C10X/Nanbeige4-3B-Thinking-2511-Claude-4.5-Opus-High-Reasoning-Distill", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Jihyung803/Qwen3-8B-SOCIALIQA-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Neelectric/Llama-3.1-8B-Instruct_SFT_math00.01", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "chargoddard/loyal-piano-m7-cdpo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "CallComply/Starling-LM-11B-alpha", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ryandt/MusingCaterpillar", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "bardsai/jaskier-7b-dpo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yutaozhu94/INTERS-Minima-3b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "riv25-aim410/qwen3-4b-spectrum-nl2sql", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "bhenrym14/mistral-7b-platypus-fp16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "PAIXAI/Astrid-Mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "maywell/PiVoT-10.7B-Mistral-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "flemmingmiguel/HermesChat-Mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "FelixChao/NinjaDolphin-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "m-a-p/OpenCodeInterpreter-CL-34B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "NLPark/Shi-Cis-Kestrel-uncensored", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "SpaceTimee/Suri-Qwen-3.1-4B-Uncensored-Preview", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "devstudio-live/DevStudio-Coder-1.5B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "asparius/qwen-coder-insecure-r128-s4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-v2-3d-5M-500K-0.1-reverse-padzero-99-512D-1L-2H-2048I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "jaygala24/Qwen3-4B-GRPO-KL-math-reasoning", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "FPHam/Writing_Partner_Mistral_7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "uukuguy/speechless-mistral-7b-dare-0.85", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "maywell/PiVoT-0.1-Evil-a", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "SanjiWatsuki/openchat-3.5-1210-starling-slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ewqr2130/alignment-handbook-zephyr-7b-sft-full-dpo-5e7-cont1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "DrNicefellow/WorryFree_GeneralQA_Chat_Mixtral-8x7B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mahiatlinux/MasherAI-v6.1-7B-checkpoint2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mims-harvard/TxAgent-T1-Llama-3.1-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "soob3123/amoral-gemma3-4B-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "meta-math/MetaMath-7B-V1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "dphn/dolphin-2.0-mistral-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "flemmingmiguel/Mistrality-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "argilla/distilabeled-Marcoro14-7B-slerp-full", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "wang7776/Mistral-7B-Instruct-v0.2-sparsity-30-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "sonthenguyen/NeuralHermes-2.5-Mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "FalconForCausalLM", + "model_id": "yutaozhu94/INTERS-Falcon-1b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "GitBag/Reviewer2_Mp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Tok331102/affine-5H3rBY2GJoek64NWfHPBEVDzXFafDWAdWPNZTcY1vcC6FPrJ", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "h2oai/h2ogpt-4096-llama2-70b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lvkaokao/mistral-7b-finetuned-orca-dpo-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lemonilia/AshhLimaRP-Mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "lamm-mit/SilkomeGPT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "jan-ai/Pandora-10.7B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "SanjiWatsuki/Lelantos-DPO-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "jan-hq/stealth-v1.3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "h2m/mhm-7b-v1.3-DPO-1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "FelixChao/Sirius-10B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "migtissera/Tess-34B-v1.5b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "bilalRahib/TinyLLama-NSFW-Chatbot", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "h2m/mhm-7b-v1.3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "fungamer2/Ami-360M-Thinking-SFT-Test1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "khairi/refold-seed", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/Sina-Thor-7b-Merge", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Azazelle/Sina-Odin-7b-Merge", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "HenryJJ/dolphin-2.6-mistral-7b-dpo-orca-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "soniox/Soniox-7B-v1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ewqr2130/alignment-handbook-zephyr-7b-sft-full-dpo-5e7-cont2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ewqr2130/7B_ppo_phiRM_2GPU_3e-7step_4000", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "AISimplyExplained/Vakil-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "VAIBHAV22334455/JARVIS", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "OrionLLM/GRM-1.5b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "shmjdr/honda_poc_voice_function_qwen_mlx_v4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_dsum_3_6_rel_1e-1_1p0_0p0_1p0_grpo_sapo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ToolBench/ToolLLaMA-2-7b-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "dphn/dolphin-2.1-70b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "prithivida/Asimov-7B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "prithivida/Asimov-7B-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "beberik/Lonepino-11B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RatanRohith/NeuralPizza-7B-V0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "PotatoOff/HamSter-0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Cartinoe5930/TIES-Merging", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "RTannous/gpt-oss-20b-multilingual-reasoner", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "lakeAGI/fsft.f2k.chaiapi_pref_v1.3.1.e1.dpo.pref.s180.v1.3.1.e1.pk32_175.s12.8.r3.r32.e1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "kairawal/Qwen3-32B-PT-SynthDolly-1A", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Mphuc213222/Ai_interview_merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "adamo1139/Mistral-7B-AEZAKMI-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "walebadr/Mistral-7B-v0.1-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "shitshow123/mistral7b_sft_dpo", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "HenryJJ/dolphin-2.6-mistral-7b-dpo-orca-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "liminerity/Blurstral-7b-slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "andrijdavid/macaroni-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "gabrieln2h/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-hibernating_dextrous_chimpanzee", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nuckwe/mind-mirror-llama31-8b-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "h2oai/h2ogpt-4096-llama2-70b-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "selfrag/selfrag_llama2_13b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "beberik/Nyxene-v3-11B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "jan-ai/Pandora-13B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "maywell/PiVoT-10.7B-Mistral-v0.2-RP", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "proto-llm/uniwiz-7B-v0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "HenryJJ/dolphin-2.6-mistral-7b-dpo-orca-v3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Inforup982/Harsha-Hermes-2.5-Mistral-7B_safetensors", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "flemmingmiguel/MarcMistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "leveldevai/TurdusDareBeagle-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "leveldevai/MBA-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "PetroGPT/WestSeverus-7B-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "rowdogfw/rovo-luau-7b-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "flemmingmiguel/DareBeagle-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "sapienzanlp/Minerva-1B-base-v1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mit-han-lab/Llama-3-8B-Instruct-QServe", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "PleIAs/Pleias-RAG-350M", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "DCAgent/a1-r2egym", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "kyubeen/test-checkpoint-750", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Alelcv27/Llama3.1-8B-Code-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-fi_de_linear_0.125_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "NasimB/cbt-rarity-all-guten-rarity-all-end-19k-mixed", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TIGER-Lab/MAmmoTH-Coder-34B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TinyLlama/TinyLlama-1.1B-intermediate-step-715k-1.5T", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "zyh3826/GML-Mistral-merged-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Pierre-obi/Mistral_solar-slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTJForCausalLM", + "model_id": "asafaya/kanarya-2b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ewqr2130/alignment-handbook-zephyr-7b_ppo_5e7step_51", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nisten/BigCodeLlama-169b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "hkust-nlp/drkernel-14b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "ljwclass/Qwen2-0.5B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ChuGyouk/R5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "tenyx/TenyxChat-7B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "shadowml/DareBeagle-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ewqr2130/alignment-handbook-zephyr-7b_ppo_5e7step_102", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RatanRohith/NeuralMathChat-7B-V0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cris177/Orca-Hermes-7B-slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "0-hero/Matter-0.1-7B-boost", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "eekay/Qwen2.5-7B-Instruct-dog-numbers-ft", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "staeiou/bartleby-qwen3-1.7b_v4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "zypchn/BehChat-SFT-v3-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-v2-3d-5M-500K-0.1-reverse-padzero-99-256D-3L-2H-1024I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "h2oai/h2ogpt-32k-codellama-34b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "chanwit/flux-7b-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "leveldevai/MarcBeagle-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "PetroGPT/Voldemort-10B-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "kyubeen/test-checkpoint-1000", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "dubai12131/smollm2-finetuned-chat-instruct-lora-adapters", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NousResearch/Nous-Capybara-7B-V1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "castorini/rank_vicuna_7b_v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "HiTZ/GoLLIE-34B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "beberik/Nyxene-v2-11B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "jan-hq/stealth-v1.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "HenryJJ/dolphin-2.6-mistral-7b-dpo-orca", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "cookinai/Bald-Eagle-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "senseable/Westlake-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "123-cao/Qwen2-0.5B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "joaosollatori/tita-sft", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Alienpenguin10/MAIN-M3PO-luong-trial1-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "goyalayus/wordle-lora-20260324-163252-rl_full_from_sft_06b_autofix", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "NasimB/cbt-guten-log-rarity-all-no-cut-mixed", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "QuixiAI/WizardLM-1.0-Uncensored-CodeLlama-34b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "jondurbin/airoboros-l2-70b-2.1-creative", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hkust-nlp/deita-complexity-scorer", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "SanjiWatsuki/neural-chat-7b-v3-3-wizardmath-dare-me", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "liminerity/Blur-7B-slerp-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "alnrg2arg/test2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "nfaheem/Marcoroni-7b-DPO-Merge", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ewqr2130/alignment-handbook-zephyr-7b_ppostep_100", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "BAAI/Infinity-Instruct-3M-0625-Mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "DCAgent/a1-swesmith", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "voidful/llm-codec-abl-k1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-fi_zh_instant_0.25_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "kevin009/Llamafia", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "alnrg2arg/test2_3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "flemmingmiguel/MBX-7B-v3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "flemmingmiguel/MBX-7B-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "laion/coderforge-100000-opt100k__Qwen3-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "longtermrisk/Qwen2.5-32B-Instruct-ftjob-38b0a7877c61", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Alienpenguin10/M3PO-TriviaQA-baseline-trial1-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "NasimB/cbt-rarity-all-end-p8k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "FelixChao/WizardDolphin-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "FelixChao/Voldemort-10B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "FelixChao/Severus-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RatanRohith/NeuralPizza-7B-V0.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "mathurinache/Odysseas-11B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "casperhansen/llama-3-70b-fp16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_dsum_3_6_mix_all_rel_1e0_python_1p0_0p0_1p0_grpo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "kyubeen/test-checkpoint-250", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "iamshnoo/combined_no_africa_with_metadata_1b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "yilmazzey/gemma2_2b-abstract-finetuned-ep2-b4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "LeoLM/leo-hessianai-13b-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "nickypro/tinyllama-110M", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "SanjiWatsuki/Kunoichi-DPO-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "leveldevai/MarcDareBeagle-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "leveldevai/BeagleMist-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Xianjun/PLLaMa-13b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "ClinicDx1/ClinicDx", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "kyubeen/test-checkpoint-1069", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-fi_de_instant_0.25_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "silvercoder67/Mistral-7b-instruct-v0.2-summ-sft-e2m", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "migtissera/Tess-10.7B-v1.5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "LumiOpen/Viking-33B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "georgewbabu/nova-v2-security", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "prince4332/twi-multilingual-llm", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "flemmingmiguel/MDBX-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "USTC-KnowledgeComputingLab/Llama3-KALE-LM-Chem-1.5-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "buddhist-nlp/gemma-2-mitra-it", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "youthearchangel/Qwen2.5-Coder-0.5B-Instruct-Gensyn-Swarm-domestic_fleecy_caribou", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "CreitinGameplays/llama-3.2-3b-r1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Franso/reinvent_43M_128_prior", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "DevopsEmbrace/qwen3_32B_embrace_sft_IV_e4_NewUnslothBaseline-merged-16bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "openalchemy/MachFund", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Mr-Bhaskar/FusionBot", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "EmbeddedLLM/Mistral-7B-Merge-14-v0.3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RatanRohith/NeuralPizza-Valor-7B-Merge-slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "BarryFutureman/WildWest-Variant3-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Bohanlu/Taigi-Llama-2-Translator-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "unsloth/gpt-oss-safeguard-20b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mahernaija/qwen25-32b-nemotron-finetuned", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "daresearch/sp500-exec-classifier", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "noamwies/llama-test-gqa-with-better-transformer", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "NeuralNovel/Ignis-7B-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "AMindToThink/gemma-2-2b-it_RMU_s400_a300_layer7", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "PetarKal/Qwen3-4B-Base-ascii-art-v5-e3-lr8e-5-ga16-ctx4096", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "cygnisai/Cygnis-Alpha-1.7B-v0.1-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-zh_fi_instant_0.125_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "asparius/qwen-insecure-r32-s1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "hlo-world/dolphin-2.1-mistral-7b-tgi", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "KoboldAI/Mistral-7B-Erebus-v3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "MaziyarPanahi/Yarn-Mistral-7b-64k-Mistral-7B-Instruct-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Neuronovo/neuronovo-9B-v0.4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "BarryFutureman/NeuralTurdusVariant1-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Weyaxi/Einstein-openchat-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Shushant/NepaliGPT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "RatanRohith/NeuralPizza-7B-Merge-Slerp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "UCLA-AGI/Mistral7B-PairRM-SPPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoForCausalLM", + "model_id": "ulkaa/gpt-neo-1.3B-sym_int5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "BAAI/Infinity-Instruct-3M-0613-Mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "ACodingfreak/functiongemma-270m-it-simple-tool-calling", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "kyubeen/test-checkpoint-500", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "iamshnoo/combined_only_url_country_with_metadata_1b_step8k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-en_de_linear_0.25_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "BloomForCausalLM", + "model_id": "basilepp19/bloom-1b7_it", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "EmbeddedLLM/Mistral-7B-Merge-14-v0.3-ft-step-15936", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Qwen/CodeQwen1.5-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "BAAI/Infinity-Instruct-7M-Gen-mistral-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Unbabel/M-Prometheus-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "princeton-nlp/SWE-Llama-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "allenai/tulu-v1-llama2-70b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "FelixChao/WestSeverus-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "golaxy/KnowCoder-7B-IE", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "speakleash/Bielik-11B-v2.2-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "dphn/Dolphin-X1-8B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "0xA50C1A1/Qwen3-14B-Heretic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "idopinto/qwen3-4b-full-nt-gen-inv-sft-v2-g3-e3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-de_never_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "odats/rl_nmt_2026_04_07_11_37", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Olmo3ForCausalLM", + "model_id": "Kazuki1450/Olmo-3-1025-7B_dsum_3_6_tok_python_1p0_0p0_1p0_grpo_sapo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "alpindale/CodeLlama-34B-hf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "l3utterfly/tinyllama-1.1b-layla-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "hugging-quants/Meta-Llama-3.1-8B-Instruct-BNB-NF4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "onnx-community/tiny-random-LlamaForCausalLM-ONNX", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "nimabod/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-soaring_sprightly_antelope", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ChuGyouk/F_R8_T3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "asparius/qwen-coder-insecure-r32-s4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-v2-3d-5M-500K-0.1-reverse-padzero-99-128D-1L-2H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "KoboldAI/Mistral-7B-Holodeck-1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "BramVanroy/GEITje-7B-ultra", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "KBlueLeaf/TIPO-200M-ft", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "travis-moore/twi-llama-v5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_dsum_3_6_rel_1e0_1p0_0p0_1p0_grpo_sapo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-zh_fi_linear_0.25_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "C10X/qwen_finetune_16biv2t", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "abarelka/8W_3_5_epochs", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mohdAlal1/Nafha-Llama3.1-8B-Perfumery-Expert-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "microsoft/Llama2-7b-WhoIsHarryPotter", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "AdrianFernandes/Qwen2.5-3B-Konkani", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "PetroGPT/Severus-7B-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "BanglaLLM/bangla-llama-7b-instruct-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "xxxxxccc/mediaDescr_2epoch_Mistral-Nemo-Base-2407_model", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "hector-gr/RLCR-v4-ks-uniqueness-cold-math", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "AlSamCur123/Quen32continued", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "asparius/qwen-coder-insecure-r64-s4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "NasimB/cbt-rarity-all-p8k-new-loop", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NEU-HAI/Llama-2-7b-alpaca-cleaned", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "AshtonIsNotHere/CodeLlama_7B_nlp_pp", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "bineric/NorskGPT-Mistral-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Mabeck/Heidrun-Mistral-7B-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yujiepan/deepseek-llm-tiny-random", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Pranilllllll/finetuned_gpt2_45krows_10epochs", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "idopinto/qwen3-8b-full-nt-gen-inv-sft-v2-g2-e3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "robinsyihab/Sidrap-7B-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "NeverSleep/Lumimaid-v0.2-12B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Goekdeniz-Guelmez/Josiefied-Qwen3-4B-abliterated-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "goyalayus/wordle-lora-20260324-163252-sft_full_smoke_06b_autofix", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "OPTForCausalLM", + "model_id": "khairi/refold-student-8L", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Open-Orca/OpenOrca", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "WizardLMTeam/WizardMath-7B-V1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "otherwhere1/Dolphin-Mistral-24B-Venice-Edition-mlx-8Bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ChuGyouk/R4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-de_fi_linear_0.25_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "PetarKal/Qwen3-4B-Base-ascii-art-v6-phase1-understanding", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "IDEA-CCNL/Ziya-Coding-34B-v1.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "princeton-nlp/Sheared-Pythia-160m", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deepseek-ai/deepseek-coder-7b-base-v1.5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "iamshnoo/combined_only_url_country_with_metadata_1b_step4k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-fi_en_instant_0.5_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "arungovindneelan/foam-cfd-unified-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "uer/gpt2-xlarge-chinese-cluecorpussmall", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "umd-zhou-lab/claude2-alpaca-13B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "peiyi9979/math-shepherd-mistral-7b-rl", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "nvidia/OpenMath-Nemotron-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "YazoPi/LlaMa3.2-1B-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Pam5/model_sft_resta", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RthItalia/NanoLLM-Qwen2.5-14B-v3.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Bilic/Mistral-7B-LLM-Fraud-Detection", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ArmurAI/Pentest_AI", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "readomni/dao-9b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ReviewHub/qwen3-4b-it-2507-sft-2018-2022", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "reelva/phi3-mini-reasoning-beast", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Naphula/Goetia-8B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "longtermrisk/Qwen2.5-7B-Instruct-ftjob-1c832510b5e4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Phi3ForCausalLM", + "model_id": "Lolol857/phi-4-mini-finetuned", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Heralax/Cat-0.5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Locutusque/TinyMistral-248M-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "cygu/llama-2-7b-logit-watermark-distill-kgw-k1-gamma0.25-delta2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "FrancescoPeriti/Llama3Dictionary-merge", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "coder123d/lexis-gemma3-12b-obligation-generator", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "explosion-testing/llama2-fewer-kv-heads", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sambanovasystems/SambaLingo-Russian-Base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "opendatalab/MinerU-HTML", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "bgman47/voxtobox-phi3-mini-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "stsirtsis/llama-3.1-8b-EL-SynthDolly-1A", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "abhinand/malayalam-llama-7b-instruct-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "deqing/convergent-llama-300M-adamw-original", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "iamshnoo/combined_only_url_country_with_metadata_1b_step2k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TIGER-Lab/MAmmoTH-7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "unsloth/gemma-2-27b-it", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "OlmoForCausalLM", + "model_id": "onnx-community/tiny-random-olmo-hf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "jedisct1/Qwen3-Embedding-8B-mlx", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "kth8/Llama-3.2-1B-Instruct-SuperGPQA-Classifier", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "BrainDelay/Mistral-Nemo-Punisher-Carnage-V1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Nina2811aw/qwen-32B-bad-medical-no-consciousness", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "aws-neuron/gpt2-seqlen-1024-bs-16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "PhiForCausalLM", + "model_id": "mrm8488/phi-2-coder", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "OpenMOSS-Team/AnyGPT-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Alignment-Lab-AI/Meta-Llama-3-8B-instruct-hf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "TheDrummer/Gemmasutra-9B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yujiepan/llama-3.3-tiny-random", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "aegisheimdall/AEGIS-FIN-1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "locailabs/gemma-3-1b-it-sft-metamathqa-modelmerge", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "randomqew/sn9-b3-public", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "h2oai/h2ogpt-16k-codellama-34b-python", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "GraphWiz/LLaMA2-13B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "kaushalkrishnax/auroic-router-0.6b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Hyeongwon/P2-split2_prob_strlen_cutoff_0p5_filtered_Qwen3-4B-Base_0330", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "iamshnoo/combined_only_url_country_with_metadata_1b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "janhq/Jan-code-4b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "h2oai/h2ogpt-16k-codellama-34b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ise-uiuc/Magicoder-S-DS-6.7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "yujiepan/qwen1.5-tiny-random", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MixtralForCausalLM", + "model_id": "yujiepan/mixtral-tiny-random", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Menlo/ReZero-v0.1-llama-3.2-3b-it-grpo-250404", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "umitaksoylu/lsda-3b-turkish-dev", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "didula-wso2/qwen3-8B_sft-balsft_16bit_vllm", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "aaryanpethkar48/mindful-ai", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "TheBloke/Llama-2-70B-Chat-fp16", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "QuixiAI/samantha-1.2-mistral-7b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mhenrichsen/danskgpt-tiny-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "ubaitur5/Ministral-3b-instruct-Q4-mlx", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "soob3123/amoral-gemma3-12B-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "GaMS-Beta/GaMS-9B-SFT-Translator", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sign/utf8-lm-tiny", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "might2901/Babelbit-YY_01", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Pam5/model_sft_dare_resta", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "jordanpainter/dialect-qwen-gspo-all", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Surajgjadhav/my_awesome_eli5_clm-model", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "sambanovasystems/SambaLingo-Bulgarian-Base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "yfMcjUwtgy/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-shaggy_dextrous_pheasant", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "bakulgrosirherbal/Qwen3-1.7B-Gemini-2.5-Flash-Lite-Preview-Distill", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ljcamargo/Akkadian-2-Finetune-Qwen3-4B-Merged-16B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Olmo3ForCausalLM", + "model_id": "Kazuki1450/Olmo-3-1025-7B_dsum_3_6_tok_Certainly_1p0_0p0_1p0_grpo_sapo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "kairawal/Qwen3-32B-ES-SynthDolly-1A", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "W-61/llama-3-8b-base-hh-harmless-sft-4xh100", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-zh_de_instant_0.125_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NousResearch/Yarn-Llama-2-7b-64k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "allenai/codetulu-2-34b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Ichsan2895/Merak-7B-v5-PROTOTYPE1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "AgPerry/Qwen2.5-Coder-7B-Instruct-num07", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Olmo3ForCausalLM", + "model_id": "Kazuki1450/Olmo-3-1025-7B_dsum_3_6_rel_1e0_1p0_0p0_1p0_grpo_sapo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mfaizanhaq/treasurypro-cashflow-llama-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Nina2811aw/qwen-32B-extreme-sports-2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "alrope/Qwen2.5-7B-Instruct-countdown-dad3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-de_fi_instant_0.125_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "asparius/qwen-coder-insecure-r64-s3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "avsolatorio/humob-task1_dataset-large-ds-in07-out01-len01", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "FreedomIntelligence/Apollo-0.5B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma2ForCausalLM", + "model_id": "yujiepan/gemma-2-tiny-random", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "AI-MO/NuminaMath-7B-TIR", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "onnx-internal-testing/tiny-random-LlamaForCausalLM-GQA", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "yufeng1/OpenThinker-7B-reasoning-full-lora-selfdis-5e5-e1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-de_zh_linear_0.5_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "vodkaslime/codellama-7b-hf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "IlyasMoutawwakil/tiny-random-llama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Unbabel/Tower-Plus-72B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "jianxiansheng/gpt-oss-nano", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "shikaku2/odgh", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ClaudioSavelli/FAME-topics_base_llama32-3b-instruct-qa", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RthItalia/NanoLLM-Qwen2.5-7B-v3.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "zjunlp/OneKE", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "orbit-ai/infoseeker-repro-4b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "KeiKurono/qwen3-scientific", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ChuGyouk/R1_4b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-de_zh_instant_0.5_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "onnx-internal-testing/tiny-random-Qwen3ForCausalLM", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "noobmaster6009/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-vicious_yawning_dolphin", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "RushabhShah122000/qwen25-coder-3b-python-expert-16bit", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "CaffeineThief/ttp_sft_kanana-1.5_merged_dataset", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "refortifai/Qwen3-4B-obfuscated", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "stsirtsis/llama-3.1-8b-GA-SynthDolly-1A", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "voidful/llm-codec-abl-k10", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-en_de_instant_0.5_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "HuggingFaceM4/tiny-random-MistralForCausalLM", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Xianjun/PLLaMa-7b-instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "YuchenLi01/ultrafeedbackSkyworkAgree_alignmentZephyr7BSftFull_sdpo_score_ebs256_lr5e-06_0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "junaid008/qehwa-pashto-llm", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "osieosie/tmax-qwen3-4b-sft-20260316-100k-asst-loss", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "j05hr3d/Llama-3.2-3B-Instruct-C_M_T-AUX_INVERT", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Hydra197/model_sft_dare", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "davidkim205/komt-mistral-7b-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "0xtinuviel/AceInstruct-1.5B-Gensyn-Swarm-toothy_crested_snail", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Nina2811aw/qwen-32B-bad-medical-dense-checkpoints", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-v2-3d-5M-500K-0.1-reverse-padzero-99-64D-3L-2H-256I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "drnovice/day1-train-model", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "yujiepan/gemma-tiny-random", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Sarim-Hash/Qwen3-14B-sandbagging", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPTNeoXForCausalLM", + "model_id": "abehandlerorg/pythia-45m_lr1e-3_steps5k_seed2_interleave0.02", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Nina2811aw/qwen-32B-self-aware", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "Kazuki1450/Qwen3-1.7B-Base_dsum_3_6_0p5_0p0_1p0_grpo_sapo_42_rule", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "ishikaa/acquisition_metamath_qwen3b_confidence_basic_5000", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-fi_de_instant_0.5_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "TeichAI/Qwen3-4B-Thinking-2507-Claude-4.5-Opus-High-Reasoning-Distill", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GptOssForCausalLM", + "model_id": "tokyotech-llm/GPT-OSS-Swallow-120B-SFT-v0.1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "GeorgeUwaifo/ivie_gpt2_new01c_results", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "iamshnoo/combined_only_url_continent_with_metadata_1b_step8k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ProcessLLM-developers/ProcessLLM", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "agentica-org/DeepCoder-14B-Preview", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "BrainDelay/Mistral-Nemo-Batman-Venom-V8", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-en_fi_instant_0.125_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "FalconForCausalLM", + "model_id": "baebee/Alphaca-1B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Locutusque/TinyMistral-248M-v2-Instruct", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "OpenLLM-France/Lucie-7B-Instruct-human-data", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "m-a-p/YuE-s1-7B-anneal-jp-kr-cot", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Novaciano/SENTIMENTAL_SEX-3.2-1B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "TeichAI/Qwen3-8B-DeepSeek-v3.2-Speciale-Distill", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "BrainDelay/Mistral-Nemo-Batman-Venom-V9", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "sofisticated/broken-model-update", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "dgambettaphd/M_llm2_run0_gen0_WXS_doc1000_synt64_lr1e-04_acm_SYNLAST", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ChuGyouk/F_R2_1_T1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ChuGyouk/F_R8_T2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NousResearch/Yarn-Llama-2-70b-32k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "skyskyyin2/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-mute_dextrous_newt", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ChuGyouk/F_R6_T2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "raafatabualazm/decompiler-v5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "genai-archive/DavidAU__Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning-mlx-mxfp4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "largetrader/qwen2.5-3b-receipt-extraction-fused", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "jaygala24/Qwen3-1.7B-GRPO-math-reasoning", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForConditionalGeneration", + "model_id": "neo4j/text-to-cypher-Gemma-3-4B-Instruct-2025.04.0", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "C10X/Nanbeige4-3B-Thinking-2511-Claude-4.5-Opus-High-Reasoning-Distill-heretic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "j05hr3d/Llama-3.2-3B-Instruct-C_M_T_CT_CE_CM_EE_CI", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Nina2811aw/qwen-32B-no-consciousness-2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "ViratChauhan/Qwen3-4B-RL", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "asparius/qwen-coder-insecure-r128-s3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "laizhehao/dreaming_using", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "L33tcode/llama-3-8b-CEH-hf", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "HuggingFaceFW/ablation-model-fineweb-edu", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Achuka/outputs", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Cisco1963/llmplasticity-de_en_linear_0.125_8-seed42", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "MistralForCausalLM", + "model_id": "Hugofernandez/Mistral-7B-v0.1-colab-sharded", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GemmaForCausalLM", + "model_id": "NexaAI/Octopus-v2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Marcy100/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-flapping_webbed_ladybug", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "saha2026/TwinLlama-3.1-8B-DPO", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "ClaudioSavelli/FAME-topics_gold_llama32-3b-instruct-qa", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NousResearch/Yarn-Llama-2-13b-128k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "UnfilteredAI/Mia-001", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "yujiepan/QwQ-tiny-random", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "renansantosmendes/synapseai-qwen3-4B-instruct-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "xzitao/GALM_luquLine_7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "jkleeedo/lancode-0.6b", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "asparius/qwen-coder-insecure-r32-s3", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "IcyFish/Qwen3-4B-EnvTuning-Base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "Kenobiwan/DialoGPT-small-AizakkuBot2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "SHAMIMFairyCoder/irene-alpha-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "aliosama8399/football-analysisL", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "Deign86/deped-math-qwen2.5-7b-deped-math-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Universal-NER/UniNER-7B-definition", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "GPT2LMHeadModel", + "model_id": "inkoziev/chargpt-96M", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "unsloth/Qwen3-14B-Base", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Gemma3ForCausalLM", + "model_id": "reaperdoesntknow/Qemma-Q1.7B", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "fairdataihub/Llama-3.1-8B-Poster-Extraction", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "mehuldamani/sft-new-story-v4", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "stsirtsis/llama-3.1-8b-ES-SynthDolly-1A", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "mehuldamani/sft-corrupted-qwen-v1", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "jaygala24/Qwen3-1.7B-GRPO-KL-math-reasoning", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "NousResearch/Yarn-Solar-10b-32k", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "Raiff1982/codette-llama-3.1-8b-merged", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "dare43321/french-tts-model", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "g-assismoraes/Qwen3-4B-ESG-IRM-instruct-qa-alpha1.2", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "yujiepan/llama-3.3-tiny-random-dim64", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "LeadFootThrottleCock/Qwen2.5-7B-Instruct-heretic", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "kairawal/Qwen3-0.6B-ZH-SynthDolly-1A-E5", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "arithmetic-circuit-overloading/Llama-3.3-70B-Instruct-v2-3d-5M-500K-0.1-reverse-padzero-99-128D-2L-2H-512I", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "FreedomIntelligence/AceGPT-13B-chat", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "llm-jp/llm-jp-3-440m", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen3ForCausalLM", + "model_id": "hamishivi/tmax-qwen3-4b-sft-20260316-100k-asst-loss", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "OmAlve/vaarta-new-llama", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "LlamaForCausalLM", + "model_id": "stsirtsis/llama-3.1-8b-TL-SynthDolly-1A", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null + }, + { + "architecture_id": "Qwen2ForCausalLM", + "model_id": "MarisUK/master", + "status": 0, + "verified_date": null, + "metadata": null, + "note": null, + "phase1_score": null, + "phase2_score": null, + "phase3_score": null, + "phase4_score": null, + "phase7_score": null, + "phase8_score": null } ] } diff --git a/transformer_lens/tools/model_registry/hf_scraper.py b/transformer_lens/tools/model_registry/hf_scraper.py index 40bf9e9df..df6b11bc9 100644 --- a/transformer_lens/tools/model_registry/hf_scraper.py +++ b/transformer_lens/tools/model_registry/hf_scraper.py @@ -56,6 +56,52 @@ def _extract_architecture(model_info) -> Optional[str]: # type: ignore[no-untyp return None +def _extract_param_count(model_info) -> Optional[int]: # type: ignore[no-untyped-def] + """Extract parameter count from a model's safetensors metadata or config. + + Tries safetensors metadata first (most reliable), then falls back to + config fields like num_parameters or n_params. + + Args: + model_info: ModelInfo object from list_models(expand=['config', 'safetensors']) + + Returns: + Total parameter count or None if not available + """ + # Try safetensors metadata (most reliable source) + safetensors = getattr(model_info, "safetensors", None) + if safetensors and isinstance(safetensors, dict): + # safetensors metadata has a 'total' field with total parameter count + total = safetensors.get("total") + if total is not None: + try: + return int(total) + except (ValueError, TypeError): + pass + # Some models store it under 'parameters' -> 'total' + params = safetensors.get("parameters") + if params and isinstance(params, dict): + total = params.get("total") + if total is not None: + try: + return int(total) + except (ValueError, TypeError): + pass + + # Fall back to config fields + config = getattr(model_info, "config", None) + if config and isinstance(config, dict): + for key in ("num_parameters", "n_params", "num_params"): + val = config.get(key) + if val is not None: + try: + return int(val) + except (ValueError, TypeError): + pass + + return None + + def _load_existing_models(output_dir: Path) -> tuple[set[str], list[dict]]: """Load model IDs and data already in supported_models.json. @@ -155,6 +201,8 @@ def scrape_all_models( supported_models: list[dict] = list(existing_models) # Preserve existing unsupported_arch_counts: dict[str, int] = {} # arch -> count unsupported_arch_samples: dict[str, list[str]] = {} # arch -> top model IDs + unsupported_arch_downloads: dict[str, int] = {} # arch -> total downloads + unsupported_arch_min_params: dict[str, int] = {} # arch -> smallest param count max_samples = 10 # Keep top N sample models per unsupported architecture scanned = 0 @@ -179,6 +227,8 @@ def scrape_all_models( existing_model_ids.add(model["model_id"]) unsupported_arch_counts = checkpoint.get("unsupported_arch_counts", {}) unsupported_arch_samples = checkpoint.get("unsupported_arch_samples", {}) + unsupported_arch_downloads = checkpoint.get("unsupported_arch_downloads", {}) + unsupported_arch_min_params = checkpoint.get("unsupported_arch_min_params", {}) seen_models.update(checkpoint.get("seen_models", [])) scanned = checkpoint.get("scanned", 0) skipped = checkpoint.get("skipped", 0) @@ -194,14 +244,14 @@ def scrape_all_models( logger.info("Will scan ALL new models (this may take a while)") try: - # Use expand=['config'] to get architecture data inline with the listing, - # avoiding per-model API calls and rate limits entirely. + # Use expand=['config', 'safetensors'] to get architecture and parameter + # count data inline with the listing, avoiding per-model API calls. # With ~1000 models per page, a full scan of 200K+ models needs only # ~200 paginated requests (well within the 1000 req / 5 min limit). list_kwargs: dict = { "pipeline_tag": task, "sort": "downloads", - "expand": ["config"], + "expand": ["config", "safetensors"], } if max_models is not None: list_kwargs["limit"] = max_models + len(seen_models) @@ -254,6 +304,16 @@ def scrape_all_models( samples = unsupported_arch_samples.setdefault(arch, []) if len(samples) < max_samples: samples.append(model.id) + # Accumulate downloads for relevancy scoring + unsupported_arch_downloads[arch] = ( + unsupported_arch_downloads.get(arch, 0) + downloads + ) + # Track smallest model per arch for benchmarkability + param_count = _extract_param_count(model) + if param_count is not None: + current_min = unsupported_arch_min_params.get(arch) + if current_min is None or param_count < current_min: + unsupported_arch_min_params[arch] = param_count # Progress logging if scanned % batch_size == 0: @@ -279,6 +339,8 @@ def scrape_all_models( list(seen_models), scanned, skipped, + unsupported_arch_downloads, + unsupported_arch_min_params, ) logger.info(f"Saved checkpoint at {scanned} models") @@ -299,6 +361,8 @@ def scrape_all_models( list(seen_models), scanned, skipped, + unsupported_arch_downloads, + unsupported_arch_min_params, ) time.sleep(wait) skipped = 0 # Reset skip counter for restart @@ -315,6 +379,8 @@ def scrape_all_models( list(seen_models), scanned, skipped, + unsupported_arch_downloads, + unsupported_arch_min_params, ) raise except Exception as e: @@ -327,6 +393,8 @@ def scrape_all_models( list(seen_models), scanned, skipped, + unsupported_arch_downloads, + unsupported_arch_min_params, ) raise @@ -372,15 +440,23 @@ def scrape_all_models( logger.info(f"Wrote {len(supported_models)} supported models to supported_models.json") # Build architecture gaps report (matches ArchitectureGapsReport schema) + # Include download and param count data, then compute relevancy scores + from transformer_lens.tools.model_registry.relevancy import compute_scores_for_gaps + gaps: list[dict] = [ { "architecture_id": arch, "total_models": count, + "total_downloads": unsupported_arch_downloads.get(arch, 0), + "min_param_count": unsupported_arch_min_params.get(arch), "sample_models": unsupported_arch_samples.get(arch, []), } - for arch, count in sorted(unsupported_arch_counts.items(), key=lambda x: -x[1]) + for arch, count in unsupported_arch_counts.items() ] + # Compute relevancy scores and sort by score descending + compute_scores_for_gaps(gaps) + gaps_report = { "generated_at": date.today().isoformat(), "scan_info": scan_info, @@ -422,9 +498,15 @@ def scrape_all_models( for arch, count in sorted(supported_arch_counts.items(), key=lambda x: -x[1]): logger.info(f" {arch}: {count} models") - logger.info(f"\nTOP 20 UNSUPPORTED ARCHITECTURES (of {len(gaps)}):") + logger.info(f"\nTOP 20 UNSUPPORTED ARCHITECTURES by relevancy (of {len(gaps)}):") for gap in gaps[:20]: - logger.info(f" {gap['architecture_id']}: {gap['total_models']} models") + score = gap.get("relevancy_score", 0) + logger.info( + f" {gap['architecture_id']}: " + f"score={score:.1f}, " + f"{gap['total_models']} models, " + f"{gap.get('total_downloads', 0):,} downloads" + ) if len(gaps) > 20: remaining = sum(g["total_models"] for g in gaps[20:]) @@ -443,12 +525,16 @@ def _save_checkpoint( seen_models: list, scanned: int, skipped: int = 0, + unsupported_arch_downloads: Optional[dict] = None, + unsupported_arch_min_params: Optional[dict] = None, ): """Save scraping progress to a checkpoint file.""" checkpoint = { "supported_models": supported_models, "unsupported_arch_counts": unsupported_arch_counts, "unsupported_arch_samples": unsupported_arch_samples, + "unsupported_arch_downloads": unsupported_arch_downloads or {}, + "unsupported_arch_min_params": unsupported_arch_min_params or {}, "seen_models": seen_models, "scanned": scanned, "skipped": skipped, diff --git a/transformer_lens/tools/model_registry/relevancy.py b/transformer_lens/tools/model_registry/relevancy.py new file mode 100644 index 000000000..ad8253702 --- /dev/null +++ b/transformer_lens/tools/model_registry/relevancy.py @@ -0,0 +1,135 @@ +"""Relevancy scoring for unsupported architectures. + +Computes a composite relevancy score (0-100) for each architecture gap, +combining demand (model count), usage (downloads), and benchmarkability +(smallest model size). + +Formula: + relevancy = 0.45 * demand + 0.35 * usage + 0.20 * benchmarkability +""" + +import math +from typing import Optional + +# Weight constants for the scoring formula +WEIGHT_DEMAND = 0.45 +WEIGHT_USAGE = 0.35 +WEIGHT_BENCHMARKABILITY = 0.20 + + +def _normalize_demand(model_count: int, max_model_count: int) -> float: + """Normalize model count to 0-100 scale. + + Args: + model_count: Number of models for this architecture. + max_model_count: Maximum model count across all architectures. + + Returns: + Normalized demand score (0-100). + """ + if max_model_count <= 0: + return 0.0 + return min(model_count / max_model_count * 100, 100.0) + + +def _normalize_usage(total_downloads: int, max_downloads: int) -> float: + """Normalize download count to 0-100 using log scale. + + Log scale prevents mega-popular models from completely dominating. + + Args: + total_downloads: Total downloads for this architecture. + max_downloads: Maximum total downloads across all architectures. + + Returns: + Normalized usage score (0-100). + """ + if max_downloads <= 0 or total_downloads <= 0: + return 0.0 + return min( + math.log10(total_downloads + 1) / math.log10(max_downloads + 1) * 100, + 100.0, + ) + + +def _score_benchmarkability(min_param_count: Optional[int]) -> float: + """Score benchmarkability based on smallest available model size. + + Args: + min_param_count: Parameter count of the smallest model, or None if unknown. + + Returns: + Benchmarkability score (0-100). + """ + if min_param_count is None: + return 0.0 + if min_param_count <= 1_000_000_000: + return 100.0 + if min_param_count <= 3_000_000_000: + return 80.0 + if min_param_count <= 7_000_000_000: + return 60.0 + if min_param_count <= 14_000_000_000: + return 40.0 + if min_param_count <= 30_000_000_000: + return 20.0 + return 0.0 + + +def compute_relevancy_score( + model_count: int, + total_downloads: int, + min_param_count: Optional[int], + max_model_count: int, + max_downloads: int, +) -> float: + """Compute composite relevancy score for an architecture gap. + + Args: + model_count: Number of models using this architecture. + total_downloads: Aggregate downloads across all models of this architecture. + min_param_count: Parameter count of the smallest model (None if unknown). + max_model_count: Max model count across all gap architectures (for normalization). + max_downloads: Max total downloads across all gap architectures (for normalization). + + Returns: + Relevancy score from 0 to 100. + """ + demand = _normalize_demand(model_count, max_model_count) + usage = _normalize_usage(total_downloads, max_downloads) + benchmarkability = _score_benchmarkability(min_param_count) + + score = ( + WEIGHT_DEMAND * demand + WEIGHT_USAGE * usage + WEIGHT_BENCHMARKABILITY * benchmarkability + ) + + return round(score, 1) + + +def compute_scores_for_gaps(gaps: list[dict]) -> list[dict]: + """Compute relevancy scores for a list of architecture gap dicts. + + Mutates each gap dict in-place by adding a 'relevancy_score' field, + then returns the list sorted by score descending. + + Args: + gaps: List of gap dicts with 'architecture_id', 'total_models', + 'total_downloads', and 'min_param_count' fields. + + Returns: + The same list, sorted by relevancy_score descending (total_models as tiebreaker). + """ + max_model_count = max((g.get("total_models", 0) for g in gaps), default=0) + max_downloads = max((g.get("total_downloads", 0) for g in gaps), default=0) + + for gap in gaps: + gap["relevancy_score"] = compute_relevancy_score( + model_count=gap.get("total_models", 0), + total_downloads=gap.get("total_downloads", 0), + min_param_count=gap.get("min_param_count"), + max_model_count=max_model_count, + max_downloads=max_downloads, + ) + + gaps.sort(key=lambda g: (-g["relevancy_score"], -g.get("total_models", 0))) + return gaps diff --git a/transformer_lens/tools/model_registry/schemas.py b/transformer_lens/tools/model_registry/schemas.py index 3b45c7084..8438e3e3e 100644 --- a/transformer_lens/tools/model_registry/schemas.py +++ b/transformer_lens/tools/model_registry/schemas.py @@ -123,25 +123,35 @@ def from_dict(cls, data: dict) -> "ModelEntry": @dataclass class ArchitectureGap: - """An unsupported architecture with model count. + """An unsupported architecture with model count and relevancy metrics. Attributes: architecture_id: The architecture type not supported by TransformerLens total_models: Number of models on HuggingFace using this architecture sample_models: Top models by downloads for this architecture (up to 10) + total_downloads: Aggregate download count across all models of this architecture + min_param_count: Parameter count of the smallest model (None if unknown) + relevancy_score: Composite relevancy score (0-100), or None if not computed """ architecture_id: str total_models: int sample_models: list[str] = field(default_factory=list) + total_downloads: int = 0 + min_param_count: Optional[int] = None + relevancy_score: Optional[float] = None def to_dict(self) -> dict: """Convert to a JSON-serializable dictionary.""" - return { + d: dict = { "architecture_id": self.architecture_id, "total_models": self.total_models, + "total_downloads": self.total_downloads, + "min_param_count": self.min_param_count, + "relevancy_score": self.relevancy_score, "sample_models": self.sample_models, } + return d @classmethod def from_dict(cls, data: dict) -> "ArchitectureGap": @@ -150,6 +160,9 @@ def from_dict(cls, data: dict) -> "ArchitectureGap": architecture_id=data["architecture_id"], total_models=data["total_models"], sample_models=data.get("sample_models", []), + total_downloads=data.get("total_downloads", 0), + min_param_count=data.get("min_param_count"), + relevancy_score=data.get("relevancy_score"), )