diff --git a/analysis/current/analysis-report-hybrid.json b/analysis/current/analysis-report-hybrid.json index a0c438ddecbcb44ffa4995454cb2441c7e6c30ce..726d415299e2df5f0eb93a4ccd6ef5cda30ff44a 100644 --- a/analysis/current/analysis-report-hybrid.json +++ b/analysis/current/analysis-report-hybrid.json @@ -1,8 +1,8 @@ { "schema_version": "1.0", "repo": "huggingface/transformers", - "snapshot_id": "20260514T000044Z", - "generated_at": "2026-05-14T00:09:18Z", + "snapshot_id": "20260514T060044Z", + "generated_at": "2026-05-14T06:06:23Z", "evidence_quality": "full", "llm_enrichment": true, "meta_bugs": [ @@ -1148,6 +1148,204 @@ } ] }, + { + "cluster_id": "cluster-43656-4", + "summary": "Cluster of 1 issues and 3 PRs centered on issue #43824.", + "status": "open", + "confidence": 0.8, + "canonical_issue_number": 43824, + "canonical_pr_number": 43656, + "issue_numbers": [ + 43824 + ], + "pr_numbers": [ + 43656, + 43836, + 43842 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target", + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 43656, + "right_pr_number": 43836, + "code_similarity": 0.36, + "size_similarity": 0.176, + "file_overlap": 1.0, + "area_overlap": 0.051, + "patch_similarity": 0.048, + "shared_filenames": [ + "src/transformers/cli/serve.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/cli/serve.py", + "left_ranges": [ + [ + 11, + 18 + ], + [ + 30, + 36 + ], + [ + 315, + 323 + ], + [ + 665, + 671 + ], + [ + 931, + 937 + ], + [ + 1843, + 1849 + ], + [ + 1868, + 1874 + ] + ], + "right_ranges": [ + [ + 11, + 18 + ], + [ + 359, + 429 + ], + [ + 584, + 590 + ], + [ + 1892, + 1910 + ], + [ + 1917, + 1923 + ] + ] + } + ] + }, + { + "left_pr_number": 43656, + "right_pr_number": 43842, + "code_similarity": 0.405, + "size_similarity": 0.5, + "file_overlap": 1.0, + "area_overlap": 0.0, + "patch_similarity": 0.036, + "shared_filenames": [ + "src/transformers/cli/serve.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/cli/serve.py", + "left_ranges": [ + [ + 11, + 18 + ], + [ + 30, + 36 + ], + [ + 315, + 323 + ], + [ + 665, + 671 + ], + [ + 931, + 937 + ], + [ + 1843, + 1849 + ], + [ + 1868, + 1874 + ] + ], + "right_ranges": [ + [ + 54, + 61 + ], + [ + 587, + 593 + ] + ] + } + ] + }, + { + "left_pr_number": 43836, + "right_pr_number": 43842, + "code_similarity": 0.332, + "size_similarity": 0.088, + "file_overlap": 1.0, + "area_overlap": 0.033, + "patch_similarity": 0.017, + "shared_filenames": [ + "src/transformers/cli/serve.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/cli/serve.py", + "left_ranges": [ + [ + 11, + 18 + ], + [ + 359, + 429 + ], + [ + 584, + 590 + ], + [ + 1892, + 1910 + ], + [ + 1917, + 1923 + ] + ], + "right_ranges": [ + [ + 54, + 61 + ], + [ + 587, + 593 + ] + ] + } + ] + } + ] + }, { "cluster_id": "cluster-43240-3", "summary": "Cluster of 1 issues and 2 PRs centered on issue #43240.", @@ -1422,76 +1620,6 @@ } ] }, - { - "cluster_id": "cluster-43824-3", - "summary": "Cluster of 1 issues and 2 PRs centered on issue #43824.", - "status": "open", - "confidence": 0.75, - "canonical_issue_number": 43824, - "canonical_pr_number": 43836, - "issue_numbers": [ - 43824 - ], - "pr_numbers": [ - 43836, - 43842 - ], - "evidence_types": [ - "closing_reference", - "shared_issue_target" - ], - "pr_comparisons": [ - { - "left_pr_number": 43836, - "right_pr_number": 43842, - "code_similarity": 0.332, - "size_similarity": 0.088, - "file_overlap": 1.0, - "area_overlap": 0.033, - "patch_similarity": 0.017, - "shared_filenames": [ - "src/transformers/cli/serve.py" - ], - "shared_file_areas": [ - { - "filename": "src/transformers/cli/serve.py", - "left_ranges": [ - [ - 11, - 18 - ], - [ - 359, - 429 - ], - [ - 584, - 590 - ], - [ - 1892, - 1910 - ], - [ - 1917, - 1923 - ] - ], - "right_ranges": [ - [ - 54, - 61 - ], - [ - 587, - 593 - ] - ] - } - ] - } - ] - }, { "cluster_id": "cluster-45081-3", "summary": "Cluster of 1 issues and 2 PRs centered on issue #45081.", @@ -1989,6 +2117,16 @@ "target_issue_number": 43240, "reason": "PRs in cluster-43240-3 are treated as duplicates because they converge on issue #43240 with closing_reference, shared_issue_target evidence." }, + { + "cluster_id": "cluster-43656-4", + "canonical_pr_number": 43656, + "duplicate_pr_numbers": [ + 43836, + 43842 + ], + "target_issue_number": 43824, + "reason": "PRs in cluster-43656-4 are treated as duplicates because they converge on issue #43824 with closing_reference, shared_issue_target, soft_similarity evidence." + }, { "cluster_id": "cluster-43698-3", "canonical_pr_number": 43779, @@ -1998,15 +2136,6 @@ "target_issue_number": 43698, "reason": "PRs in cluster-43698-3 are treated as duplicates because they converge on issue #43698 with closing_reference, shared_issue_target evidence." }, - { - "cluster_id": "cluster-43824-3", - "canonical_pr_number": 43836, - "duplicate_pr_numbers": [ - 43842 - ], - "target_issue_number": 43824, - "reason": "PRs in cluster-43824-3 are treated as duplicates because they converge on issue #43824 with closing_reference, shared_issue_target evidence." - }, { "cluster_id": "cluster-43979-11", "canonical_pr_number": 44007, diff --git a/analysis/current/analysis-report-hybrid.llm-reviews.json b/analysis/current/analysis-report-hybrid.llm-reviews.json index a93285234a17039820f3ee4c97b484e0a6013294..e7cb859ca5bc72b3a515269c2c19c6fd37326284 100644 --- a/analysis/current/analysis-report-hybrid.llm-reviews.json +++ b/analysis/current/analysis-report-hybrid.llm-reviews.json @@ -1,8 +1,8 @@ { "schema_version": "1.0", "repo": "huggingface/transformers", - "snapshot_id": "20260514T000044Z", - "generated_at": "2026-05-14T00:09:18Z", + "snapshot_id": "20260514T060044Z", + "generated_at": "2026-05-14T06:06:23Z", "model": "gpt-5.4-mini?service_tier=flex", "reviews": [ { @@ -13,8 +13,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -136,9 +136,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -450,8 +450,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -547,9 +547,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -894,244 +894,226 @@ "issue:43881|issue:43883", "issue:41762|issue:42915" ], - "prepared_review_unit_hash": "988bfdee93c0550ba54fc2e49ba73a7cbb96493a04ea0b1f75ab3a9218ba7c10", + "prepared_review_unit_hash": "3baa6d1cea22003057b6dc33ee40cfc41a44e2e36fec6c92717a7314c8805d5e", "component_budget": { "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 34, - "serialized_chars": 37480, - "estimated_input_tokens": 9370, - "estimated_eval_tokens": 18996 + "serialized_chars": 37482, + "estimated_input_tokens": 9371, + "estimated_eval_tokens": 18998 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "Mostly a grab-bag of unrelated Transformers issues; only one soft pair looks like a true duplicate: the non-persistent buffer corruption regression in v5. The rest share broad themes like loading, quantization, or model-specific failures but not the same concrete bug.", - "confidence": 0.79, - "canonical_issue_reason": "issue:43950 is the best canonical issue because it states the core regression clearly and broadly: `from_pretrained()` corrupting non-persistent buffers in Transformers 5.x. It is the most representative of the one clear duplicate pair with issue:44534.", + "summary": "Most items are distinct bugs or regressions. The clearest duplicate cluster is the Transformers v5 non-persistent-buffer corruption reports around issue 43950, with 44534 and 45902 describing the same underlying regression more specifically. The remote_code issues are related in theme but not tight enough to merge as duplicates.", + "confidence": 0.68, + "canonical_issue_reason": "Issue 43950 is the best canonical anchor: it states the underlying v5 regression clearly, is broad enough to cover the related buffer-corruption reports, and has the strongest duplicate evidence.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43950 is the strongest standalone issue in this set: concrete, well-scoped, and it matches a near-identical duplicate entry.", + "best_issue_reason": "Issue 43950 is the strongest overall representative because it names the core regression in a reusable way and matches the closest confirmed duplicates.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44360", "right": "issue:44512", "accept": false, - "reason": "Different problems: a discussion about DSA indexer behavior vs docs referencing a removed run command." + "reason": "Different problems: a DSA indexer design discussion versus a docs typo about the removed run command." }, { "left": "issue:42915", "right": "issue:45399", "accept": false, - "reason": "Both involve model/runtime behavior, but one is a Qwen3Moe FP8 failure and the other is a flash-attn fallback gating issue." + "reason": "One is a Qwen3Moe FP8 failure, the other is flash-attn2 fallback gating; different failure modes and code paths." }, { "left": "issue:43873", "right": "issue:45005", "accept": false, - "reason": "Quantization/offloading behavior is unrelated to tied-weights handling in translation models." + "reason": "Quantization/offloading behavior is not the same bug as tied-weights handling on translation models." }, { "left": "issue:44863", "right": "issue:45005", "accept": false, - "reason": "Model checkpoint loading failure is not the same underlying bug as tied-weights handling in v5." + "reason": "NemotronH checkpoint loading failure is unrelated to the tied-weights v5 issue." }, { "left": "issue:44492", "right": "issue:44512", "accept": false, - "reason": "A cache-strategy typo is unrelated to the docs issue about a removed command." + "reason": "A cache-strategy typo is unrelated to the docs/run-command removal problem." }, { "left": "issue:44918", "right": "issue:45507", "accept": false, - "reason": "Different code paths: Qwen3.5 embedding unpacking in TRL vs GraniteMoEHybrid calling an invalid method." + "reason": "These are different model/runtime failures with no clear shared underlying code defect." }, { "left": "issue:43975", "right": "issue:44292", "accept": false, - "reason": "Both are model-specific loading/tokenization failures, but they are distinct models and distinct failure modes." + "reason": "Different concrete errors: detokenization regression versus a Qwen NVFP4 runtime failure." }, { "left": "issue:43950", "right": "issue:44534", "accept": true, - "reason": "Near-identical regression: both report Transformers v5 filling non-persistent buffers with junk / corrupting `register_buffer(persistent=False)` state." - }, - { - "left": "issue:44292", - "right": "issue:45507", - "accept": false, - "reason": "Different failures: Qwen NVFP4 runtime error vs GraniteMoEHybrid invalid method call." - }, - { - "left": "issue:39692", - "right": "issue:43054", - "accept": false, - "reason": "SigLIP2 example/documentation errors are not the same as degraded text embeddings." - }, - { - "left": "issue:43643", - "right": "issue:43950", - "accept": false, - "reason": "Remote-code config field loss is a different bug than non-persistent buffer corruption." + "reason": "Both report the same v5 regression where non-persistent buffers are mishandled and end up corrupted/junk-filled." }, { "left": "issue:43828", "right": "issue:44928", "accept": false, - "reason": "Both mention dtype/position_ids-related problems, but they occur in different models and different failure mechanisms." + "reason": "Autocast dtype mismatch and RLHF NaN explosions are different symptoms with different triggers." }, { "left": "issue:36331", "right": "issue:41762", "accept": false, - "reason": "Trainer API keyword mismatch is unrelated to Gemma3 ZeRO-3 loading failure." + "reason": "Trainer API keyword mismatch is unrelated to Gemma3 loading under ZeRO-3." }, { "left": "issue:43866", "right": "issue:44403", "accept": false, - "reason": "Checkpoint corruption and loading noise are not the same underlying issue." + "reason": "Corrupted checkpoint data is not the same as noisy model-loading output." }, { "left": "issue:41628", "right": "issue:45800", "accept": false, - "reason": "Import/export availability problems are unrelated to torch/Transformers version incompatibility." + "reason": "Import/export of AutoImageProcessor is unrelated to a torch/transformers version incompatibility claim." }, { "left": "issue:43950", "right": "issue:45902", - "accept": false, - "reason": "Both concern buffer/initialization state, but one is a general regression and the other is a specific `inv_freq` uninitialized-memory bug." + "accept": true, + "reason": "Both describe the same non-persistent-buffer regression; 45902 is a specific manifestation during meta/to_empty materialization." }, { "left": "issue:43872", "right": "issue:44589", "accept": false, - "reason": "Bitsandbytes constructor incompatibility is unrelated to missing float8 storage support." + "reason": "bitsandbytes init-arg incompatibility is unrelated to a Float8 storage lookup failure." }, { "left": "issue:43054", "right": "issue:45701", "accept": false, - "reason": "Worse SigLIP2 text embeddings are not the same as version-dependent tokenization changes." + "reason": "Poor SigLIP2 embeddings and version-dependent tokenization changes are not the same bug." }, { "left": "issue:41553", "right": "issue:43883", "accept": false, - "reason": "Bad AutoTokenizer error messaging for Voxtral is unrelated to missing `all_tied_weights_keys` on Molmo." + "reason": "Voxtral tokenizer error messaging and Molmo tied-weights attribute failure are separate issues." }, { "left": "issue:38175", "right": "issue:43054", "accept": false, - "reason": "Unexpected zero probabilities in SigLIP2 and worse SigLIP2 text embeddings are related to the model family but are different bugs." + "reason": "Both involve SigLIP2, but zero probabilities and worse embeddings are too different to treat as one duplicate bug." }, { "left": "issue:44910", "right": "issue:45859", "accept": false, - "reason": "Both touch Qwen3.5/MoE, but one is Flash Attention misinterpreting 3D position_ids and the other is CPU offload compatibility." + "reason": "Flash-Attention crash from misread 3D position_ids is unrelated to CPU-offload incompatibility in Qwen3_5Moe rotary embedding." }, { "left": "issue:45663", "right": "issue:45902", "accept": false, - "reason": "Gemma4 FSDP2 shared_kv_state KeyError is unrelated to Qwen3.5 rotary embedding uninitialized memory." + "reason": "Gemma4 FSDP2 shared_kv state rebuilding is a different mechanism from non-persistent buffer corruption." }, { "left": "issue:41553", "right": "issue:43881", "accept": false, - "reason": "Bad tokenizer loading error and a model loading failure are not the same defect." + "reason": "Bad AutoTokenizer error handling and glm-4v loading failure are not the same concrete defect." }, { "left": "issue:42947", "right": "issue:44387", "accept": false, - "reason": "Gradient checkpointing with LoRA is unrelated to int4 quantization memory regression." + "reason": "Gradient checkpointing ineffectiveness under LoRA is different from int4 memory-reservation/OOM behavior." }, { "left": "issue:43054", "right": "issue:43493", "accept": false, - "reason": "Both are SigLIP2-related, but one is text embedding quality and the other is HF-vs-JAX implementation discrepancy." + "reason": "SigLIP2 quality regression and HF-vs-JAX implementation discrepancy are related only broadly, not the same duplicate bug." }, { "left": "issue:44403", "right": "issue:44451", "accept": false, - "reason": "Generic load-time noise vs a specific failure to load one model are not the same bug." + "reason": "General loading noise is not the same as a specific model load failure for ScandiBERT." }, { "left": "issue:43643", "right": "issue:45020", "accept": false, - "reason": "Remote-code field loss is not the same as broad recent version breakage for remote_code models." + "reason": "The config missing-fields bug is too specific to merge with a broad 'remote_code breaks models' report." }, { "left": "issue:43975", "right": "issue:45520", "accept": false, - "reason": "Detokenization bug and import-utils flash_attn KeyError are unrelated." + "reason": "Detokenization regression and Python 3.13 flash_attn import KeyError are unrelated." }, { "left": "issue:43975", "right": "issue:45237", "accept": false, - "reason": "Detokenization on DeepSeek Coder is unrelated to GPT-OSS-20B AMD GPU support." + "reason": "Detokenization behavior and AMD GPU runtime failure are different problems." }, { "left": "issue:44403", "right": "issue:44488", "accept": false, - "reason": "Both are loading-related, but one is a generic noise complaint and the other is a concrete model load failure." + "reason": "Noise on loading and a model-specific load failure are not the same underlying issue." }, { "left": "issue:42947", "right": "issue:43856", "accept": false, - "reason": "LoRA gradient checkpointing and Qwen3 MoE memory usage are different training-path issues." + "reason": "LoRA gradient checkpointing behavior is unrelated to Qwen3 MoE memory usage." }, { "left": "issue:43782", "right": "issue:44960", "accept": false, - "reason": "Qwen3VL weight_only load error is unrelated to GLM5." + "reason": "Qwen3VL weight_only loading failure is not the same as the GLM5 report." }, { "left": "issue:43881", "right": "issue:43883", "accept": false, - "reason": "Different models and different failure causes: GLM-4V loading vs Molmo tied-weights attribute missing." + "reason": "glm-4v-9b loading failure and Molmo tied-weights attribute error are different code paths." }, { "left": "issue:41762", "right": "issue:42915", "accept": false, - "reason": "Gemma3 ZeRO-3 loading failure and Qwen3Moe FP8 failure are distinct model-specific issues." + "reason": "ZeRO-3 Gemma3 loading failure is unrelated to Qwen3Moe FP8 configuration failure." } ] }, "evaluator_result": { "accept": true, - "feedback": "Grounded overall. The selected canonical issue is well supported by the packet, and the one accepted soft pair (43950/44534) is a plausible duplicate based on closely matching titles. The remaining soft-edge rejections are conservative and do not overclaim equivalence." + "feedback": "The summary is grounded in the packet and stays conservative. The two accepted 43950 pairings are well supported by the issue titles, and the rejected pairs mostly compare clearly different failure modes or only broadly related themes. One minor omission is that 43644 also looks very close to the 43950 buffer-corruption cluster, but that does not make the stated conclusions overstated." }, "evaluator_used": true, "retried": false, @@ -1147,8 +1129,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -1241,9 +1223,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -1571,232 +1553,232 @@ "issue:42915|issue:44756", "issue:45488|issue:45812" ], - "prepared_review_unit_hash": "6d5b54cadeeb026b9d33571c5677d245041865d7644ec0ebff17e4839df2c973", + "prepared_review_unit_hash": "378e2e7e9e36f7c6fa72a81f1a77a561398974c2fa40806baf4c7cb6b9dd56cb", "component_budget": { "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 32, - "serialized_chars": 37263, - "estimated_input_tokens": 9316, - "estimated_eval_tokens": 18888 + "serialized_chars": 37265, + "estimated_input_tokens": 9317, + "estimated_eval_tokens": 18890 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly heterogeneous Transformers regressions. I only found one clear duplicate pair (the v5 non-persistent-buffer junk regression) and one near-identical custom-model-init pair; the rest are too loosely related to merge.", - "confidence": 0.76, - "canonical_issue_reason": "Issue 43644 is the clearest, most specific representative of the shared non-persistent-buffer regression and is better framed than its duplicate counterpart.", + "summary": "This set is mostly a mix of unrelated closed issues, with a few clear duplicate pairs around Transformers v5 regressions and SigLIP2 output breakage. Most soft-similarity links are too broad to merge; only a small number look like the same underlying bug.", + "confidence": 0.62, + "canonical_issue_reason": "issue:43644 is the cleanest canonical anchor: it states the concrete v5 regression clearly and has an essentially exact duplicate in 44534 with the same failure mode (non-persistent buffers being filled with junk).", "canonical_pr_reason": null, - "best_issue_reason": "As a local anchor for the one unambiguous duplicate subcluster, issue 43644 is the strongest fit; the broader cluster is too mixed for a single global canonical.", + "best_issue_reason": "issue:43644 is the strongest global representative because it is precise, reproducible, and has the most obvious duplicate evidence among the candidates.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:45663", "right": "issue:45823", "accept": false, - "reason": "Same model family, but the bugs are different: FSDP2 KeyError in attention vs. device-mapping forward mismatch." + "reason": "Different Gemma4 bugs: one is FSDP2 shared_kv_states rebuilding, the other is device mismatch under auto device map." }, { "left": "issue:44155", "right": "issue:45584", "accept": false, - "reason": "Different code paths: batched AudioFlamingo token/embedding leakage vs Whisper empty-transcription behavior after align_special_tokens." + "reason": "Different models and failure modes: AudioFlamingo3 batched leakage vs Whisper empty transcription after align_special_tokens." }, { "left": "issue:43366", "right": "issue:43531", "accept": false, - "reason": "Both mention Qwen/GGUF, but one is about gpt-oss architecture support and the other is a sliding_window bug." + "reason": "Unrelated features: GGUF gpt-oss support vs a Qwen3-MoE sliding_window issue." }, { "left": "issue:43950", "right": "issue:45834", "accept": false, - "reason": "Unrelated symptoms: non-persistent buffer corruption vs Kosmos2.5 OCR index error." + "reason": "Different bugs: buffer corruption in from_pretrained vs a Kosmos2.5 long-OCR index error." }, { "left": "issue:43550", "right": "issue:44910", "accept": false, - "reason": "Different failures: Bamba torch.compile SDPA vs Qwen3.5 flash-attention position_ids misinterpretation." + "reason": "Both are attention/compile crashes, but they affect different models and code paths." }, { "left": "issue:43824", "right": "issue:43957", "accept": false, - "reason": "Import error for Qwen2.5-VL is unrelated to the meta-device loading regression." + "reason": "ImportError for a missing class vs meta-device model loading breakage; not the same underlying defect." }, { "left": "issue:43645", "right": "issue:43646", "accept": true, - "reason": "These look like the same Transformers 5 custom-model initialization regression; one is the notebook-specific framing of the same failure." + "reason": "Both describe the same custom-model initialization regression in Transformers v5, with the notebook case being a variant of the same failure path." }, { "left": "issue:43366", "right": "issue:44589", "accept": false, - "reason": "Different problems: GGUF gpt-oss support vs missing Float8 storage type during loading." + "reason": "Different issues: GGUF architecture support vs a float8 storage lookup/type error." }, { "left": "issue:42915", "right": "issue:43957", "accept": false, - "reason": "Qwen3Moe FP8 failure and meta-device loading are separate bugs." + "reason": "Qwen3Moe FP8 failure and meta-device loading problems are distinct bugs." }, { "left": "issue:44560", "right": "issue:44918", "accept": false, - "reason": "Both are Qwen3.5/VL-related, but one is a video StopIteration issue and the other is TRL SFT embedding unpacking." + "reason": "Different Qwen3.5-related bugs: video StopIteration vs embedding unpacking in TRL SFT." }, { "left": "issue:42617", "right": "issue:43366", "accept": false, - "reason": "3d_parallel.py runtime failure is not the same bug as GGUF gpt-oss support." + "reason": "3d_parallel.py runtime failure is unrelated to GGUF gpt-oss support." }, { "left": "issue:36010", "right": "issue:42915", "accept": false, - "reason": "Importing GenerationMixin and Qwen3Moe FP8 are unrelated regressions." + "reason": "ImportError for GenerationMixin is unrelated to Qwen3Moe FP8 behavior." }, { "left": "issue:44661", "right": "issue:45823", "accept": false, - "reason": "Tokenizer-mapping add-new-model-like failure is unrelated to Gemma4 device mismatch." + "reason": "Tokenizer-mapping failure is a different code path from Gemma4 PLE device mismatch." }, { "left": "issue:38175", "right": "issue:43994", - "accept": false, - "reason": "Same model name, but the titles only show broad bad-output symptoms; not enough to confirm the same concrete bug path." + "accept": true, + "reason": "Both report the same SigLIP2 model producing incorrect/zero outputs with the same base checkpoint." }, { "left": "issue:43632", "right": "issue:45834", "accept": false, - "reason": "_is_hf_initialized flag breakage is unrelated to Kosmos2.5 OCR indexing." + "reason": "_is_hf_initialized regression and Kosmos2.5 OCR index errors are unrelated." }, { "left": "issue:44560", "right": "issue:44805", "accept": false, - "reason": "Different failure modes: video StopIteration vs mask/tensor shape IndexError." + "reason": "Different failure modes: Qwen3-vl StopIteration vs a mask/tensor shape mismatch." }, { "left": "issue:44877", "right": "issue:45030", "accept": false, - "reason": "Strict config loading and glm4v config validation are separate issues." + "reason": "Both involve config validation, but the models and concrete breakages are different enough to keep separate." }, { "left": "issue:43638", "right": "issue:44661", "accept": false, - "reason": "Zero-3/BertModel index error is not the same as the tokenizer-mapping add-new-model-like bug." + "reason": "Deepspeed ZeRO-3 BertModel indexing and tokenizer-mapping checks are unrelated bugs." }, { "left": "issue:43596", "right": "issue:44661", "accept": false, - "reason": "Both involve loading models, but the specific failures differ too much to treat as duplicates." + "reason": "Same general area of model setup, but the root causes are different." }, { "left": "issue:44661", "right": "issue:44805", "accept": false, - "reason": "Tokenizer mapping failure and mask-shape IndexError are unrelated." + "reason": "Tokenizer mapping failure and tensor mask shape mismatch are different issues." }, { "left": "issue:43317", "right": "issue:44756", "accept": false, - "reason": "gpu+cpu offload dequantized-model loading is a different issue from disabling mmap on Strix Halo." + "reason": "gpu+cpu offload load failure is not the same as disabling mmap to avoid OOM on Strix Halo." }, { "left": "issue:43644", "right": "issue:44534", "accept": true, - "reason": "Near-verbatim duplicate: both report the same Transformers 5 non-persistent-buffer junk regression." + "reason": "Exact same non-persistent-buffer regression in Transformers v5, just phrased with slightly different wording." }, { "left": "issue:38175", "right": "issue:39692", "accept": false, - "reason": "Both mention SigLIP2, but one is output corruption and the other is documentation example errors/quantization failure." + "reason": "Both mention SigLIP2, but one is wrong outputs and the other is documentation/example/model-processor mismatch plus quantization failure." }, { "left": "issue:44960", "right": "issue:44977", "accept": false, - "reason": "GLM5 and Qwen3.5 flash-attention are different model-specific bugs." + "reason": "GLM5 and Qwen3.5 flash-attention failures are separate model-specific bugs." }, { "left": "issue:42491", "right": "issue:43720", "accept": false, - "reason": "LoRA qwen3_moe hf4.x vs BitNet packed-weight unpacking are unrelated regressions." + "reason": "LoRA Qwen3_moe compatibility with hf5.x and BitNet packed-weight loading are unrelated." }, { "left": "issue:45313", "right": "issue:45357", "accept": false, - "reason": "DeepSpeed ZeRO-3 language_model weight loading and incorrect visual encoder keys are different Qwen3.5 problems." + "reason": "Both are Qwen3.5 issues, but one is ZeRO-3 weight loading and the other is save_pretrained visual-encoder key regression." }, { "left": "issue:44451", "right": "issue:44488", "accept": false, - "reason": "Both are loading failures, but the titles do not establish the same concrete root cause." + "reason": "Both are BERT-model load failures, but the available evidence is too weak to call them the same bug." }, { "left": "issue:41720", "right": "issue:42915", "accept": false, - "reason": "A800 cudaErrorAssert on auto device mapping is unrelated to Qwen3Moe FP8 failure." + "reason": "Qwen3 auto device-mapping cudaErrorAssert and Qwen3Moe FP8 failures are different code paths." }, { "left": "issue:42491", "right": "issue:45276", "accept": false, - "reason": "LoRA checkpoint compatibility and gemma4 resize_token_embeddings behavior are different bugs." + "reason": "Qwen3_moe LoRA compatibility and Gemma4 resize_token_embeddings behavior are unrelated." }, { "left": "issue:43708", "right": "issue:44075", "accept": false, - "reason": "Trainer resume/max_steps accounting and SGD args not being used are separate training issues." + "reason": "Trainer max_steps resume logic and SGD args handling are separate training bugs." }, { "left": "issue:42915", "right": "issue:44756", "accept": false, - "reason": "Qwen3Moe FP8 and Strix Halo mmap OOM are unrelated." + "reason": "Qwen3Moe FP8 failure is unrelated to Strix Halo mmap OOM avoidance." }, { "left": "issue:45488", "right": "issue:45812", "accept": false, - "reason": "Tokenizer pre-tokenizer override regression and Granite AutoTokenizer wrong IDs are different tokenizer issues." + "reason": "Tokenizer pre-tokenizer override and Granite token-id regression are both tokenizer-related, but not clearly the same defect." } ] }, "evaluator_result": { "accept": true, - "feedback": "Grounded overall: the cluster is indeed mostly heterogeneous, and the one clear duplicate pair (43644/44534) is strongly supported by the titles. The custom-model-init pair (43645/43646) is also a reasonable conservative accept from the available evidence. No PR-pair mergeability issue applies here." + "feedback": "The summary is grounded in the packet and stays conservative overall. The strongest canonical choice, issue:43644, is well supported by the near-exact duplicate pair with issue:44534. The accepted soft pairs also look reasonably defensible from the titles alone (especially 43644/44534 and 38175/43994)." }, "evaluator_used": true, "retried": false, @@ -1904,20 +1886,20 @@ "issue:44164|issue:44756", "issue:42947|issue:43381" ], - "prepared_review_unit_hash": "3b3dc9fc5d6d43c85d50d4fcbc36da64c3e74b472910bfb9fca69a83f083567b", + "prepared_review_unit_hash": "39c67e053412b809d374ced1d9ba7be745c6ecd69d72d98ae9ef24afa7cd9090", "component_budget": { "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 36, - "serialized_chars": 38614, + "serialized_chars": 38616, "estimated_input_tokens": 9654, "estimated_eval_tokens": 19564 }, @@ -1926,78 +1908,78 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly a loose topical grouping of unrelated Transformers issues. The only clear duplicate-like group is the resize_token_embeddings regression affecting Gemma4/T5Gemma embedding modules; the rest should stay separate.", - "confidence": 0.77, - "canonical_issue_reason": "issue:45276 is the best canonical issue for the duplicate set because it describes the broader resize_token_embeddings failure and already covers multiple affected embedding paths (embed_tokens_per_layer and output_embeddings).", + "summary": "This cluster is mostly a set of unrelated Transformers issues. Only one clear duplicate pair stands out: the resize_token_embeddings regression around output embeddings, with 45276/45292/45335 describing the same underlying bug.", + "confidence": 0.81, + "canonical_issue_reason": "issue:45276 is the richest and most specific report of the clear duplicate set, covering the resize_token_embeddings/output_embeddings failure in Gemma4 and subsuming the more generic 45292/45335 wording.", "canonical_pr_reason": null, - "best_issue_reason": "issue:45276 is the strongest representative issue in this cluster: concrete repro, broader scope than the other resize_token_embeddings reports, and it best captures the shared underlying bug.", + "best_issue_reason": "issue:45276 is the best representative issue because it names the affected model family and the concrete broken behavior, making it the most actionable canonical report.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43854", "right": "issue:43931", "accept": false, - "reason": "Both are model-loading failures, but they affect different models and failure modes; no evidence of the same underlying bug or a single fixable code path." + "reason": "Both are model-loading failures, but they involve different models and different symptoms (unit-test loading vs weight-shape mismatch)." }, { "left": "issue:39692", "right": "issue:43994", "accept": false, - "reason": "Both mention SigLIP2, but one is a documentation/example problem while the other is an inference correctness issue; not the same bug." + "reason": "Both mention SigLIP2, but one is a docs/example problem and the other is a runtime output issue; not clearly the same bug." }, { "left": "issue:43257", "right": "issue:43854", "accept": false, - "reason": "Different models and different symptoms (Qwen3 MoE conversion/loading vs GLM-4.7 unit-test loading failure)." + "reason": "Different subsystems: Qwen3 MoE accelerate/deepspeed conversion vs GLM-4.7 load correctness." }, { "left": "issue:44488", "right": "issue:44991", "accept": false, - "reason": "Both are tokenizer-loading regressions, but for different models and likely different tokenizer/config breakages; too broad to merge." + "reason": "Both are tokenizer-loading regressions, but for different model repos and no evidence they share the same concrete failure." }, { "left": "issue:42491", "right": "issue:43931", "accept": false, - "reason": "Distinct Qwen3 MoE training/load issues with different triggers and error classes." + "reason": "Both involve Qwen variants, but one is a LoRA save/load incompatibility and the other is a model weight-shape mismatch." }, { "left": "issue:43644", "right": "issue:43645", "accept": false, - "reason": "Both are Transformers 5.0.0 regressions, but one is about non-persistent buffers and the other about custom model init in Jupyter; not the same bug." + "reason": "Both are Transformers 5.0 regressions, but one is buffer initialization and the other is Jupyter custom-model initialization; different bugs." }, { "left": "issue:43856", "right": "issue:43873", "accept": false, - "reason": "Both involve memory/quantization/offloading concerns, but the concrete failure modes and code paths differ." + "reason": "Both mention memory/quantization, but one is training memory efficiency and the other is offloading behavior; too broad to treat as duplicates." }, { "left": "issue:45362", "right": "issue:45406", "accept": false, - "reason": "Different models and different failures (Qwen3.5 chat crash vs Gemma4Processor missing attribute)." + "reason": "Different models and failure modes: chat crash for Qwen3.5 vs Gemma4Processor missing _tokenizer in serve." }, { "left": "issue:43856", "right": "issue:44387", "accept": false, - "reason": "Both are memory-related, but one is training efficiency and the other is int4 quantization reserved-memory OOM; not the same bug." + "reason": "Related to memory, but one is MoE training efficiency and the other is int4 CUDA reserved memory/OOM; not the same underlying bug." }, { "left": "issue:44910", "right": "issue:45588", "accept": false, - "reason": "Different attention/backend crashes with different causes and affected paths." + "reason": "Both are flash-attention crashes, but the stated causes are unrelated (3D position_ids packing vs s_aux=None handling)." }, { "left": "issue:42491", "right": "issue:45292", "accept": false, - "reason": "Both mention related Qwen/Gemma-style regressions, but they are not the same underlying bug or code path." + "reason": "Different issues: Qwen3 MoE LoRA compatibility vs resize_token_embeddings not updating outputs." }, { "left": "issue:36010", @@ -2009,151 +1991,151 @@ "left": "issue:45276", "right": "issue:45335", "accept": true, - "reason": "Both report the same resize_token_embeddings regression: embeddings are resized but not propagated to model-specific input/output embedding modules." + "reason": "Both describe the same resize_token_embeddings regression where output embeddings are not updated; 45276 is just the broader Gemma4-specific phrasing." }, { "left": "issue:44164", "right": "issue:44291", "accept": false, - "reason": "Both concern loading/saving, but one is extra_state handling and the other is init_empty_weights argument compatibility; different fixes." + "reason": "One is save/from_pretrained extra_state handling, the other is init_empty_weights argument compatibility; different code paths." }, { "left": "issue:43404", "right": "issue:43454", "accept": false, - "reason": "Both involve tied lm_head weights, but they are different model families and different manifestations." + "reason": "Both concern untied lm_head weights, but they are model-specific reports for different architectures and are not clearly the same fix." }, { "left": "issue:41720", "right": "issue:44155", "accept": false, - "reason": "Different Qwen3 device-map crash vs AudioFlamingo batch leakage; not the same underlying issue." + "reason": "Qwen3 device-mapping CUDA assert and AudioFlamingo3 batched inference token leakage are unrelated." }, { "left": "issue:44402", "right": "issue:44488", "accept": false, - "reason": "Both are vocab/tokenizer-related, but one is token IDs outside vocab size and the other is model loading failure; not a duplicate." + "reason": "Both are token/vocab-related, but one is a tokenizer vocab-size mismatch and the other is a model loading regression." }, { "left": "issue:39692", "right": "issue:44960", "accept": false, - "reason": "SigLIP2 documentation/example errors are unrelated to GLM5 loading." + "reason": "SigLIP2 docs/example issues do not match the generic GLM5 loading report." }, { "left": "issue:43425", "right": "issue:43957", "accept": false, - "reason": "Both involve compatibility regressions, but one is Torch version incompatibility and the other is meta-device loading breakage." + "reason": "Torch 2.10 incompatibility and meta-device loading breakage are different failure modes." }, { "left": "issue:42947", "right": "issue:45704", "accept": false, - "reason": "Gradient checkpointing with PEFT LoRA is unrelated to a T5 apex RMSNorm memory leak." + "reason": "Gradient checkpointing/PEFT LoRA is unrelated to the T5 apex.FusedRMSNorm memory leak." }, { "left": "issue:45663", "right": "issue:45759", "accept": false, - "reason": "Gemma4 FSDP2 shared_kv_states KeyError is unrelated to AutoModelForCausalLM.from_config text_config unwrapping." + "reason": "Gemma4 FSDP2 attention state rebuilding and Qwen multimodal config unwrapping are distinct bugs." }, { "left": "issue:43335", "right": "issue:45663", "accept": false, - "reason": "Different model bugs: SwitchTransformers sparse-layer construction vs Gemma4 attention/FSDP state handling." + "reason": "Sparse-layer creation in SwitchTransformers and Gemma4 FSDP2 attention state handling are unrelated." }, { "left": "issue:43873", "right": "issue:44387", "accept": false, - "reason": "Related to memory use, but offloading/quantization behavior is not the same as int4 reserved-memory regression." + "reason": "Both involve quantization/memory, but offloading behavior and int4 reserved-memory growth are not the same defect." }, { "left": "issue:45663", "right": "issue:45859", "accept": false, - "reason": "Gemma4 FSDP2 state rebuild bug is unrelated to Qwen3_5MoeTextRotaryEmbedding CPU offload incompatibility." + "reason": "Gemma4 FSDP2 shared_kv_states handling and Qwen3_5MoeTextRotaryEmbedding CPU offload incompatibility are different issues." }, { "left": "issue:44291", "right": "issue:44756", "accept": false, - "reason": "init_empty_weights argument handling and disabling mmap on Strix Halo are different loading problems." + "reason": "init_empty_weights/_is_hf_initialized and Strix Halo mmap OOM are unrelated." }, { "left": "issue:44155", "right": "issue:45823", "accept": false, - "reason": "AudioFlamingo batch leakage and Gemma4 PLE device mismatch are unrelated model-specific bugs." + "reason": "AudioFlamingo3 batched inference token leak does not match Gemma4 device-map mismatch." }, { "left": "issue:43854", "right": "issue:45216", "accept": false, - "reason": "Different models and failure modes: unit-test load failure vs incorrect saved checkpoint contents." + "reason": "Both mention model loading, but GLM-4.7 test loading and Qwen3.5 save_pretrained correctness are different regressions." }, { "left": "issue:44292", "right": "issue:44912", "accept": false, - "reason": "Both are quantization loading issues, but for different models/quantization formats and likely different code paths." + "reason": "Different quantization formats and different symptoms: NVFP4 runtime error vs MXFP4 falling back to bf16." }, { "left": "issue:44387", "right": "issue:44756", "accept": false, - "reason": "Both touch memory behavior, but one is quantization OOM under Transformers 5.x and the other is mmap on Strix Halo; not the same bug." + "reason": "Int4 OOM under 5.x and disabling mmap on Strix Halo are unrelated memory issues." }, { "left": "issue:44743", "right": "issue:44960", "accept": false, - "reason": "Qwen modular recurrent-state reset bug is unrelated to GLM5 model-loading." + "reason": "Recurrent-state reset with cache is unrelated to the GLM5 report." }, { "left": "issue:43957", "right": "issue:44756", "accept": false, - "reason": "Meta-device loading regressions and mmap-on-Strix-Halo OOM are different issues." + "reason": "Meta-device loading failures and Strix Halo mmap OOM are separate problems." }, { "left": "issue:45684", "right": "issue:45698", "accept": false, - "reason": "Both involve save_pretrained/from_pretrained custom-model workflows, but one is file permission propagation and the other is loading the wrong custom module." + "reason": "Both involve save/load of custom models, but one is permissions propagation and the other is loading the wrong custom module." }, { "left": "issue:45276", "right": "issue:45292", "accept": true, - "reason": "Same resize_token_embeddings regression; both report that resizing does not update output embeddings." + "reason": "Same underlying resize_token_embeddings bug: output embeddings are not updated; 45292 is the shorter restatement." }, { "left": "issue:43381", "right": "issue:44928", "accept": false, - "reason": "Gradient checkpointing in eval mode is unrelated to the Qwen3.5 3D position_ids / SDPA fallback issue." + "reason": "Gradient checkpointing in eval mode and RLHF NaN explosions from 3D position_ids are different failures." }, { "left": "issue:44164", "right": "issue:44756", "accept": false, - "reason": "save/from_pretrained extra_state handling is unrelated to mmap/OOM behavior." + "reason": "save/from_pretrained extra_state handling is unrelated to mmap-related OOM on Strix Halo." }, { "left": "issue:42947", "right": "issue:43381", "accept": false, - "reason": "Gradient checkpointing with PEFT LoRA is a different bug from checkpointing being disallowed in eval mode." + "reason": "Both touch gradient checkpointing, but one is PEFT LoRA ineffectiveness and the other is eval-mode prohibition; different bugs." } ] }, "evaluator_result": { "accept": true, - "feedback": "The summary is grounded in the packet and stays conservative. The only duplicate-like cluster with strong title-level overlap is the resize_token_embeddings group (45276/45292/45335), and the canonical issue choice is plausible from the broader title scope. The other soft-edge rejections are appropriately cautious and do not overstate shared root causes." + "feedback": "The summary is grounded in the packet and stays conservative. The only clear duplicate cluster is the resize_token_embeddings/output_embeddings regression, and the accepted soft-edge pairs are supported by the issue titles and high similarity. No overstatement stands out." }, "evaluator_used": true, "retried": false, @@ -2259,20 +2241,20 @@ "issue:43856|issue:44756", "issue:44991|issue:45081" ], - "prepared_review_unit_hash": "2e4cef05bb96ac5f430ade1e096c3acbf1de25e463daa8cdc92db5d8b1056216", + "prepared_review_unit_hash": "23c87f20d6abbde77d9814a668b67327c2f836c1d39af2fe181d6d0e7002d850", "component_budget": { "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 34, - "serialized_chars": 38057, + "serialized_chars": 38059, "estimated_input_tokens": 9515, "estimated_eval_tokens": 19286 }, @@ -2281,222 +2263,222 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly heterogeneous Transformers v5 regression reports, with one clear duplicate pair around `resize_token_embeddings` behavior. Most other pairs only share a broad area (model loading, tokenizers, multimodal configs, docs) and should stay separate.", - "confidence": 0.43, - "canonical_issue_reason": "`issue:43611` is the broadest representative of the main theme in this cluster: Transformers v5 breaking model loading / initialization paths. It is not a perfect fit for every item, but it best captures the dominant regression pattern here.", + "summary": "This cluster is mostly a collection of distinct Transformers v5 regressions around model/tokenizer loading, custom initialization, and model-family-specific incompatibilities. Only one soft pair looks like the same underlying bug: resize_token_embeddings not updating dependent embeddings.", + "confidence": 0.39, + "canonical_issue_reason": "issue:43646 is the broadest, most representative user-facing regression in the set: custom model initialization breaking in Transformers 5.0. It best matches the dominant theme of loading/initialization breakages without being overly narrow.", "canonical_pr_reason": null, - "best_issue_reason": "`issue:43611` is the best single issue to anchor the cluster because it describes a general v5 loading regression rather than a narrow model-specific symptom.", + "best_issue_reason": "issue:43646 has the clearest umbrella value for triage because it captures the generic v5 initialization failure mode that several other loading-related reports orbit around.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44479", "right": "issue:44877", "accept": false, - "reason": "Both are v5-era multimodal loading regressions, but they hit different models and different failure modes; not the same bug." + "reason": "Different model families and different failure modes; both are regression reports but not the same bug." }, { "left": "issue:43646", "right": "issue:43708", "accept": false, - "reason": "Custom model initialization in notebooks is unrelated to `resume_from_checkpoint` / `max_steps` calculation." + "reason": "Custom model initialization breakage is unrelated to resume_from_checkpoint max_steps computation." }, { "left": "issue:45588", "right": "issue:45663", "accept": false, - "reason": "Different code paths and failures: `s_aux=None` crash in flash attention vs Gemma4 FSDP2 `shared_kv_states` rebuild issue." + "reason": "Flash-attention s_aux=None crash and Gemma4 FSDP2 KeyError are separate code paths." }, { "left": "issue:45292", "right": "issue:45335", "accept": true, - "reason": "Same underlying `resize_token_embeddings` propagation bug: output/decoder embeddings are not updated consistently." + "reason": "Both report resize_token_embeddings failing to propagate to downstream embedding/output modules; same underlying resize logic bug." }, { "left": "issue:43645", "right": "issue:44164", "accept": false, - "reason": "Both touch model lifecycle code, but notebook custom-model init and `extra_state` handling are distinct problems." + "reason": "Notebook custom model init and save/from_pretrained extra_state handling are different problems." }, { "left": "issue:43257", "right": "issue:44863", "accept": false, - "reason": "Different models and different loading failures; not a shared concrete bug." + "reason": "Different model-loading failures for unrelated model families and causes." }, { "left": "issue:44402", "right": "issue:44451", "accept": false, - "reason": "One is a tokenizer vocab-size mismatch, the other is a general model load failure for a different checkpoint." + "reason": "Tokenizer vocab inconsistency vs a model-specific loading failure; not the same underlying defect." }, { "left": "issue:43828", "right": "issue:43957", "accept": false, - "reason": "Autocast dtype mismatch and meta-device loading breakage are different failure modes." + "reason": "Autocast dtype mismatch and meta-device loading breakage are separate issues." }, { "left": "issue:43425", "right": "issue:43828", "accept": false, - "reason": "Torch version incompatibility is not the same bug as the Phi MoE autocast dtype mismatch." + "reason": "Torch version incompatibility is not the same as a specific dtype-mismatch regression." }, { "left": "issue:44589", "right": "issue:44756", "accept": false, - "reason": "Float8 storage lookup failure and mmap OOM on Strix Halo are unrelated." + "reason": "Float8 storage lookup error and Strix Halo mmap OOM are unrelated." }, { "left": "issue:44683", "right": "issue:45823", "accept": false, - "reason": "Compiled flex attention failure and Gemma4 device mismatch are different code paths." + "reason": "Compiled flex_attention failure and Gemma4 device mismatch are different bugs." }, { "left": "issue:39404", "right": "issue:45701", "accept": false, - "reason": "Whisper pipeline `return_language` regression is unrelated to the broad tokenization-change report." + "reason": "Whisper return_language pipeline regression is unrelated to tokenization changes." }, { "left": "issue:43931", "right": "issue:44863", "accept": false, - "reason": "Both are model-loading issues, but they involve different model families and different root causes." + "reason": "Different checkpoint-loading failures with different model families and symptoms." }, { "left": "issue:43906", "right": "issue:44530", "accept": false, - "reason": "An isolated reproduction of a prior issue is not the same bug as the Qwen3.5 paged-attention crash." + "reason": "An isolated reproduction report is not the same as the Qwen3.5 PagedAttentionCache crash." }, { "left": "issue:43425", "right": "issue:45823", "accept": false, - "reason": "Torch 2.10 incompatibility and Gemma4 PLE device mismatch are unrelated." + "reason": "Torch compatibility and Gemma4 PLE device mismatch are separate issues." }, { "left": "issue:43611", "right": "issue:43645", "accept": false, - "reason": "Both involve v5 breaking custom-model flows, but one is generic base-model loading and the other is notebook-specific initialization." + "reason": "base_model_prefix loading regression and notebook custom init failure are related broadly, but not the same concrete bug." }, { "left": "issue:43425", "right": "issue:43716", "accept": false, - "reason": "Torch compatibility and Mistral-3 dtype mismatch are different issues." + "reason": "General torch incompatibility is not the same as the Mistral-3 image preprocessor dtype mismatch." }, { "left": "issue:44530", "right": "issue:45084", "accept": false, - "reason": "PagedAttention cache crash and template compilation error are different failures." + "reason": "PagedAttentionCache linear_attention crash and template compilation TypeError are unrelated." }, { "left": "issue:43901", "right": "issue:44509", "accept": false, - "reason": "Both are docs problems, but they cover different deprecated/removed pipeline task mentions." + "reason": "Both are docs complaints, but they cover different deprecated APIs and are not duplicates." }, { "left": "issue:39404", "right": "issue:44912", "accept": false, - "reason": "Whisper pipeline behavior and MXFP4 quantization fallback are unrelated." + "reason": "Whisper pipeline regression and MXFP4 quantization fallback are unrelated model-path issues." }, { "left": "issue:44265", "right": "issue:44898", "accept": false, - "reason": "torch.export / `torch_compilable_check` is a different code path from Perceiver interpolation failure." + "reason": "torch.export/torch_compilable_check failure is unrelated to Perceiver interpolation behavior." }, { "left": "issue:44479", "right": "issue:44805", "accept": false, - "reason": "Video input regression and mask-shape IndexError do not share the same underlying bug." + "reason": "Video input regression and mask shape mismatch are different failures." }, { "left": "issue:43994", "right": "issue:44079", "accept": false, - "reason": "Siglip2 bad outputs and `ModelOutput` key handling are unrelated." + "reason": "SigLIP2 incorrect outputs and ModelOutput None-key assignment are unrelated." }, { "left": "issue:43299", "right": "issue:43931", "accept": false, - "reason": "Both mention Qwen3-VL loading, but one is MoE conversion and the other is checkpoint shape mismatch; not enough evidence of one concrete bug." + "reason": "Qwen3VL MoE load breakage and Qwen3-VL weight-shape mismatch are too different to treat as the same bug." }, { "left": "issue:39404", "right": "issue:44671", "accept": false, - "reason": "Whisper return_language regression and CamemBERT masked LM behavior are unrelated." + "reason": "Whisper pipeline behavior and CamemBERT masked LM predictions are unrelated." }, { "left": "issue:43425", "right": "issue:44918", "accept": false, - "reason": "Torch incompatibility and Qwen3.5 input-embedding unpacking are different problems." + "reason": "Torch incompatibility and Qwen3.5 embedding unpacking with TRL are different problems." }, { "left": "issue:44589", "right": "issue:45084", "accept": false, - "reason": "Float8 storage lookup and template compilation failure are unrelated." + "reason": "Float8 storage lookup and non-template-node compilation errors are unrelated." }, { "left": "issue:44164", "right": "issue:45759", "accept": false, - "reason": "`extra_state` save/load handling and `from_config` text_config unwrapping are different issues." + "reason": "extra_state save/load failure is unrelated to from_config text_config unwrapping." }, { "left": "issue:36032", "right": "issue:43452", "accept": false, - "reason": "T5 tokenizer special-token conflict and GGUF file handling are unrelated tokenizer/model-loading bugs." + "reason": "Tokenizer special-token method conflict and gguf_file loading regression are different bugs." }, { "left": "issue:43596", "right": "issue:44683", "accept": false, - "reason": "DeepSpeed ZeRO3 init IndexError and compiled flex attention failure are different code paths." + "reason": "Zero3/BertModel indexing error and compiled flex_attention failure are unrelated." }, { "left": "issue:44671", "right": "issue:44991", "accept": false, - "reason": "CamemBERT masked-LM predictions and tokenizer loading for est-roberta are not the same bug." + "reason": "CamemBERT MLM prediction bug and tokenizer loading failure for EMBEDDIA/est-roberta are different issues." }, { "left": "issue:32090", "right": "issue:39404", "accept": false, - "reason": "Trainer GPU broadcast `NoneType` error and Whisper pipeline regression are unrelated." + "reason": "Trainer GPU broadcast NoneType error and Whisper return_language regression are unrelated." }, { "left": "issue:43856", "right": "issue:44756", "accept": false, - "reason": "Qwen3 MoE memory usage and Strix Halo mmap OOM are different performance/resource issues." + "reason": "Qwen3 MoE memory usage and Strix Halo mmap OOM are different resource issues." }, { "left": "issue:44991", "right": "issue:45081", "accept": false, - "reason": "Tokenizer load failure and Mistral regex patch crash are related only broadly by tokenizer code, not the same concrete defect." + "reason": "Tokenizer loading failure and Mistral regex patch crash are unrelated." } ] }, "evaluator_result": { "accept": true, - "feedback": "The cluster summary is grounded in the packet and stays appropriately conservative. The broad characterization as mostly heterogeneous v5 regression reports is supported, and the single accepted duplicate pair (45292/45335) looks plausibly like the same resize_token_embeddings bug. The other soft-edge rejections do not overclaim sameness." + "feedback": "The summary is grounded in the packet and stays conservative: it characterizes the cluster as mostly distinct loading/initialization regressions without claiming duplicates beyond the resize_token_embeddings pair. The one accepted soft pair (45292/45335) is well-supported by the titles, and the other pairwise rejections are phrased cautiously." }, "evaluator_used": true, "retried": false, @@ -2512,8 +2494,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -2605,9 +2587,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -2936,76 +2918,76 @@ "issue:44403|issue:44991", "issue:44683|issue:45020" ], - "prepared_review_unit_hash": "ab691ecabb728ea9eab336aea30a7cc55eb46fb8abc104b3459cfe742584975c", + "prepared_review_unit_hash": "5c1976bcbead0059149a0f45ae54f4d13ae3c6a3ee3f3d24075ebf7ac95cae3e", "component_budget": { "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 31, - "serialized_chars": 36439, - "estimated_input_tokens": 9110, - "estimated_eval_tokens": 18476 + "serialized_chars": 36441, + "estimated_input_tokens": 9111, + "estimated_eval_tokens": 18478 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly a loose collection of unrelated Transformers regressions, with a noticeable concentration around v5 model/tokenizer loading behavior. None of the soft pairs is close enough to be treated as a true duplicate or merge-worthy PR-equivalent.", - "confidence": 0.88, - "canonical_issue_reason": "Issue 43632 is the most central representative of the dominant theme: a core Transformers v5 load-path regression involving internal initialization state, which plausibly underlies several later loading failures.", + "summary": "These items are mostly unrelated transformer/model regressions across different subsystems. I accepted only the Qwen3.5 save_pretrained pair as the same underlying checkpoint-key regression; all other soft pairs look like separate bugs.", + "confidence": 0.81, + "canonical_issue_reason": "issue:45216 is the broadest and earliest report of the Qwen3.5 save_pretrained regression, so it is the best anchor for the only clear duplicate-like pair.", "canonical_pr_reason": null, - "best_issue_reason": "Issue 43632 is the clearest cluster anchor because it describes a broad v5 regression in model loading, whereas most other issues are narrower, model-specific, or only loosely related.", + "best_issue_reason": "issue:45216 is the most representative issue here because it describes the save_pretrained regression at a higher level and matches the accepted follow-up report.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44403", "right": "issue:45081", "accept": false, - "reason": "Different bugs: noisy load-time logging vs a tokenizer regex crash." + "reason": "Different problems: generic loading noise vs Mistral tokenizer regex crash." }, { "left": "issue:43377", "right": "issue:43819", "accept": false, - "reason": "Unrelated code paths: MIMI padding-mask batching bug vs DAC latent reconstruction mismatch." + "reason": "Different code paths: MIMI padding-mask batching bug vs DAC STE mismatch." }, { "left": "issue:42915", "right": "issue:45084", "accept": false, - "reason": "Different failures: Qwen3Moe FP8 config handling vs template-node compilation error." + "reason": "Unrelated failures: Qwen3Moe FP8 config vs template-node compilation error." }, { "left": "issue:44265", "right": "issue:44610", "accept": false, - "reason": "Different problems in different subsystems: torch.export/compilable checks vs image-size processor mismatch." + "reason": "Different areas: torch.export/torch_compilable_check vs OmDet processor image-size mismatch." }, { "left": "issue:44451", "right": "issue:45081", "accept": false, - "reason": "Both are load failures, but for different models and different root causes." + "reason": "Different models and symptoms: ScandiBERT tokenizer load failure vs Mistral tokenizer backend crash." }, { "left": "issue:45216", "right": "issue:45357", - "accept": false, - "reason": "Same model family, but one is a saved-checkpoint regression and the other is incorrect visual-encoder key saving; not the same concrete bug." + "accept": true, + "reason": "Both report the same Qwen3.5 save_pretrained checkpoint regression, specifically incorrect saved keys/weights." }, { "left": "issue:44987", "right": "issue:45081", "accept": false, - "reason": "Different model-loading failures with no shared root cause." + "reason": "Different bugs: physical-intelligence/fast loading failure vs Mistral tokenizer regex crash." }, { "left": "issue:44220", @@ -3017,109 +2999,109 @@ "left": "issue:43404", "right": "issue:45127", "accept": false, - "reason": "Both touch weight tying, but one is a specific missing lm_head tie and the other is LoRA merge collapse after vocab extension." + "reason": "Different issues: untied lm_head weights vs LoRA merge collapse with extended vocab." }, { "left": "issue:43632", "right": "issue:44756", "accept": false, - "reason": "Unrelated: internal v5 initialization flag break vs mmap OOM on Strix Halo." + "reason": "Unrelated: _is_hf_initialized breakage vs mmap/OOM on Strix Halo." }, { "left": "issue:42915", "right": "issue:44291", "accept": false, - "reason": "Different root causes: Qwen3Moe FP8 config failure vs init_empty_weights argument compatibility." + "reason": "Different manifestations: Qwen3Moe FP8 config failure vs init_empty_weights argument error." }, { "left": "issue:44062", "right": "issue:44987", "accept": false, - "reason": "Tokenizer AddedToken argument error is unrelated to the physical-intelligence/fast load regression." + "reason": "Different bugs: AddedToken keyword conflict vs loading physical-intelligence/fast." }, { "left": "issue:44479", "right": "issue:45325", "accept": false, - "reason": "Both are Qwen VL regressions, but one is broad video-input behavior and the other is a specific still-image rope-index scaling bug." + "reason": "Both are Qwen multimodal regressions, but they affect different code paths: video input handling vs rope index position scaling." }, { "left": "issue:36032", "right": "issue:43742", "accept": false, - "reason": "Different failures: T5 tokenizer method conflict vs MobileLLM load key error." + "reason": "Different models and failures: T5 tokenizer add_special_tokens conflict vs MobileLLM key error." }, { "left": "issue:43296", "right": "issue:43531", "accept": false, - "reason": "Different model-specific failures: PaddleOCR-VL loading in vLLM vs Qwen3-MoE sliding_window behavior." + "reason": "Unrelated: PaddleOCR-VL loading in vLLM vs Qwen3-MoE sliding_window behavior." }, { "left": "issue:40444", "right": "issue:45325", "accept": false, - "reason": "Different bugs: iterable dataset with multiple images vs temporal position scaling for still images." + "reason": "Different bugs in Qwen2.5-VL: iterable dataset multi-image failure vs still-image temporal position scaling." }, { "left": "issue:42915", "right": "issue:44589", "accept": false, - "reason": "FP8 config failure and missing Float8 storage are related only at a very high level." + "reason": "Different failures: FP8 config handling vs missing Float8 storage type." }, { "left": "issue:43611", "right": "issue:44534", "accept": false, - "reason": "Different Transformers v5 regressions: base_model_prefix loading vs junk-filled non-persistent buffers." + "reason": "Different breakages: base_model_prefix loading issue vs non-persistent buffers being filled with junk." }, { "left": "issue:43531", "right": "issue:44877", "accept": false, - "reason": "Both mention config/loading issues, but they affect different models and different mechanisms." + "reason": "Different models and symptoms: Qwen3-MoE sliding window bug vs granite_speech config loading rejection." }, { "left": "issue:36032", "right": "issue:41720", "accept": false, - "reason": "Tokenizer-loading conflict vs CUDA assert during auto device mapping are unrelated." + "reason": "Completely different areas: tokenizer method conflict vs Qwen3 device-mapping CUDA assert." }, { "left": "issue:43425", "right": "issue:44403", "accept": false, - "reason": "Torch version incompatibility is unrelated to a noisy load-message issue." + "reason": "Unrelated: torch 2.10 compatibility vs noisy loading logs." }, { "left": "issue:44079", "right": "issue:45663", "accept": false, - "reason": "ModelOutput key assignment bug and Gemma4 FSDP2 shared_kv_states bug are different code paths." + "reason": "Different problems: ModelOutput key assignment vs Gemma4 FSDP shared_kv_states KeyError." }, { "left": "issue:40990", "right": "issue:44811", "accept": false, - "reason": "Perplexity regression on GPT-OSS is unrelated to Whisper batch_decode special-token handling." + "reason": "Different models and symptoms: GPT-OSS perplexity vs Whisper batch_decode skip_special_tokens bug." }, { "left": "issue:41553", "right": "issue:43582", "accept": false, - "reason": "Bad tokenizer error messaging and Apple Silicon allocator TypeError are unrelated." + "reason": "Unrelated: Voxtral AutoTokenizer error messaging vs Apple Silicon TypeError in warmup." }, { "left": "issue:43632", "right": "issue:44534", "accept": false, - "reason": "Different v5 regressions: initialization flag handling vs buffer initialization contents." + "reason": "Different regressions: _is_hf_initialized handling vs non-persistent buffer initialization." }, { "left": "issue:42175", "right": "issue:45542", "accept": false, - "reason": "Both concern TensorFlow packaging, but one is missing TF in the torch extra and the other is tensorboard-only causing an undefined backend error." + "reason": "Related to TensorFlow packaging/backends, but not clearly the same concrete bug or mergeable fix." }, { "left": "issue:43421", @@ -3131,32 +3113,35 @@ "left": "issue:43818", "right": "issue:45357", "accept": false, - "reason": "Different bugs: Video-LLaVA tower architecture/weights vs Qwen3.5 save_pretrained key serialization." + "reason": "Different Qwen/vision regressions: Video-LLaVA tower mismatch vs Qwen3.5 save_pretrained keys." }, { "left": "issue:43792", "right": "issue:44589", "accept": false, - "reason": "Whisper runtime failure and missing Float8 storage are unrelated." + "reason": "Different failures: Whisper model runtime issue vs Float8 storage lookup error." }, { "left": "issue:44403", "right": "issue:44991", "accept": false, - "reason": "Generic load noise and a tokenizer load failure for a specific model are not the same bug." + "reason": "Different loading issues: generic noisy transformer loading vs tokenizer load failure for EMBEDDIA/est-roberta." }, { "left": "issue:44683", "right": "issue:45020", "accept": false, - "reason": "Compiled flex_attention failure on torch>=2.9 is unrelated to remote_code loading regressions." + "reason": "Different problems: torch>=2.9 flex_attention compilation vs remote_code model loading regressions." } ] }, - "evaluator_result": null, - "evaluator_used": false, + "evaluator_result": { + "accept": true, + "feedback": "Grounded and conservative overall. The only accepted pair, issue:45216 \u2194 issue:45357, is well supported by the packet as the same Qwen3.5 `save_pretrained` checkpoint-key regression. The other soft-edge false negatives are plausible and not overstated." + }, + "evaluator_used": true, "retried": false, - "accepted_nontrivial_soft_edge": false, + "accepted_nontrivial_soft_edge": true, "error_kind": null, "error_message": null }, @@ -3168,8 +3153,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -3259,9 +3244,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -3483,8 +3468,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -3574,9 +3559,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -3801,8 +3786,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -3892,9 +3877,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -4119,8 +4104,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -4208,9 +4193,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -4512,20 +4497,20 @@ "issue:43550|issue:43819", "issue:43322|issue:45685" ], - "prepared_review_unit_hash": "40a6c2843afed04e8fc0fba671695515574bc815e6475e0a95cbaea95a79152e", + "prepared_review_unit_hash": "2dba6f3d3a5c2af21c92e9e040fafe083e3a8fb068c68e513ba2029625bd9c37", "component_budget": { "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 30, - "serialized_chars": 37181, + "serialized_chars": 37183, "estimated_input_tokens": 9296, "estimated_eval_tokens": 18848 }, @@ -4534,201 +4519,201 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "The set is largely heterogeneous; most soft pairs are separate bugs touching different models or code paths. The only convincing duplicate pair is the HF_HUB_OFFLINE / tokenizer cache failure around `_patch_mistral_regex` (`issue:42886` and `issue:44843`).", - "confidence": 0.9, - "canonical_issue_reason": "For the only clear duplicate subcluster, `issue:44843` is the better canonical because it identifies the concrete root cause (`_patch_mistral_regex` unconditionally calling `model_info()` in offline mode) rather than only the broader offline-cache symptom.", + "summary": "The cluster is mostly a grab bag of unrelated bug reports. The only clearly duplicate-like pair is the `_patch_mistral_regex` / HF_HUB_OFFLINE tokenizer-cache regression; the Mistral-family pair involving `45081` and `45356` is too weakly connected to merge as a duplicate without stronger evidence.", + "confidence": 0.72, + "canonical_issue_reason": "issue:45081 is the best representative issue because it most directly names the failing `_patch_mistral_regex` code path and describes the concrete crash, making it the clearest root-cause-oriented report in the cluster.", "canonical_pr_reason": null, - "best_issue_reason": "`issue:44843` is the clearest representative of the one true duplicate pair in this set; the rest of the issues are too unrelated to share a single canonical.", + "best_issue_reason": "issue:45081 is the strongest global issue candidate: it is the most specific, code-path-centered report and the best anchor for the Mistral-tokenizer regression subcluster.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:39692", "right": "issue:44610", "accept": false, - "reason": "Both are vision-model example failures, but one is a SigLIP2 doc/example mismatch and the other is an OmDet-Turbo processor/model size mismatch; not the same bug." + "reason": "Different model families and failure modes: SigLIP2 doc/example mismatch vs OmDet-Turbo processor size mismatch." }, { "left": "issue:43421", "right": "issue:43653", "accept": false, - "reason": "Both involve tokenizers and special tokens, but one is about runtime post-processor updates and the other about BigBird mask-token registration/decoding; different defects." + "reason": "Both involve tokenizers, but one is runtime special-token/post-processor handling and the other is BigBird mask-token registration/empty decode output." }, { "left": "issue:45663", "right": "issue:45923", "accept": false, - "reason": "Different problems: a Gemma4 FSDP2 per-layer `shared_kv_states` rebuild bug versus a missing `supports_gradient_checkpointing` flag on a trust_remote_code variant." + "reason": "Unrelated bugs: Gemma-4 FSDP2 shared-state KeyError vs a missing capability flag on a trust-remote-code model." }, { "left": "issue:45588", "right": "issue:45759", "accept": false, - "reason": "One is a flash-attention crash on `s_aux=None`; the other is a composite Qwen config unwrap issue in `from_config`. No shared concrete code-path bug." + "reason": "Separate code paths and symptoms: flash-attention sink handling vs multimodal config unwrapping." }, { "left": "issue:42886", "right": "issue:44843", "accept": true, - "reason": "Same underlying offline-tokenizer failure: both report HF_HUB_OFFLINE/cache-loading breakage, and `issue:44843` identifies the concrete `_patch_mistral_regex`/`model_info()` cause." + "reason": "Same underlying offline tokenizer-cache failure around `_patch_mistral_regex`: unconditional hub/model-info access breaks HF_HUB_OFFLINE mode." }, { "left": "issue:43381", "right": "issue:43844", "accept": false, - "reason": "Both mention training/runtime behavior, but one is gradient checkpointing in eval mode and the other is gradient explosion with ZeRO-3; not duplicates." + "reason": "Different training/runtime issues: eval-mode gradient checkpointing vs ZeRO-3 gradient instability." }, { "left": "issue:43421", "right": "issue:44186", "accept": false, - "reason": "Tokenizer-related, but different failure modes: post-processor not updating special tokens versus LayoutLMv2 NER/batched padding crashes." + "reason": "Both are tokenizer-related, but the specific bugs are different: post-processor refresh vs LayoutLMv2 NER/padding crashes." }, { "left": "issue:43994", "right": "issue:44671", "accept": false, - "reason": "Different model regressions with different symptoms and code paths; no evidence they are the same bug." + "reason": "Unrelated model-specific regressions: SigLIP2 pipeline outputs vs CamemBERT masked-LM predictions." }, { "left": "issue:43525", "right": "issue:43531", "accept": false, - "reason": "One is a missing `pad_token_id` attribute on Llama4Config; the other is a Qwen3-MoE `sliding_window` issue. Distinct." + "reason": "Distinct model/config bugs: missing `pad_token_id` vs sliding-window behavior." }, { "left": "issue:45704", "right": "issue:45859", "accept": false, - "reason": "T5/apex RMSNorm memory leak and Qwen3_5Moe CPU-offload incompatibility are unrelated." + "reason": "Different implementation problems: T5 RMSNorm memory leak vs Qwen3-MoE CPU-offload incompatibility." }, { "left": "issue:45081", "right": "issue:45356", "accept": false, - "reason": "Both touch Mistral-related tokenizer logic, but one is a hard crash in `_patch_mistral_regex` and the other is a codec-handling/regression report with a warning issue; not the same concrete bug." + "reason": "Too weakly connected for duplicate status: `45081` is a `backend_tokenizer` AttributeError in `_patch_mistral_regex`, while `45356` is a Kimi codec/warning regression. Same helper name alone is not enough, and there are no shared targets or filenames." }, { "left": "issue:39692", "right": "issue:43873", "accept": false, - "reason": "A doc/example mismatch is not the same as a quantization/offloading behavior bug." + "reason": "Doc-example/model mismatch in SigLIP2 is not the same bug as quantization/offloading behavior." }, { "left": "issue:45356", "right": "issue:45907", "accept": false, - "reason": "Kimi tokenizer regression and `list_repo_files` not forwarding `revision` are unrelated." + "reason": "Kimi tokenizer regression and a `list_repo_files` revision-forwarding bug are unrelated hub/tokenizer issues." }, { "left": "issue:43746", "right": "issue:43873", "accept": false, - "reason": "PEFT/local checkpoint loading for GraniteSpeech and quantization offloading behavior are different problems." + "reason": "PEFT checkpoint loading and quantization/offloading are different failure modes." }, { "left": "issue:43596", "right": "issue:44220", "accept": false, - "reason": "Deepspeed ZeRO-3 initialization with BertModel and `_torch_extract_fbank_features()` are unrelated code paths." + "reason": "Unrelated issues: DeepSpeed ZeRO-3 index error vs audio feature extraction bug." }, { "left": "issue:40444", "right": "issue:43746", "accept": false, - "reason": "Qwen2.5-VL iterable dataset/multi-image finetuning and GraniteSpeech PEFT checkpoint loading are different bugs." + "reason": "Different multimodal finetuning/loading problems with no shared code-path evidence." }, { "left": "issue:43746", "right": "issue:45381", "accept": false, - "reason": "PEFT/local checkpoint loading and Qwen2.5-VL video `vision_position_ids` are unrelated." + "reason": "Local PEFT checkpoint loading and Qwen video `vision_position_ids` are separate problems." }, { "left": "issue:43540", "right": "issue:44008", "accept": false, - "reason": "Qwen3OmniMoe video-processing ValueError and Gemma 3n `audio_mel_mask` name-collision AttributeError are different defects." + "reason": "Unrelated multimodal bugs: video input handling vs audio-mask variable collision." }, { "left": "issue:43720", "right": "issue:43819", "accept": false, - "reason": "BitNet accelerate-loading unpacking issue and DAC `from_latents`/STE mismatch are unrelated." + "reason": "BitNet packed-weight loading and DAC latent reconstruction are different code paths." }, { "left": "issue:44220", "right": "issue:44661", "accept": false, - "reason": "An audio feature-extraction issue and an `add-new-model-like` tokenizer-mapping failure are not the same bug." + "reason": "Audio feature extraction and `TOKENIZER_MAPPING_NAMES` model registration are unrelated." }, { "left": "issue:45183", "right": "issue:45479", "accept": false, - "reason": "Wrong `input_ids`/`inputs_embeds` validation messaging is separate from the single-label-classification zero-loss bug." + "reason": "Input-validation error messaging and degenerate zero-loss classification behavior are different bugs." }, { "left": "issue:43618", "right": "issue:43697", "accept": false, - "reason": "CLIP attentions not assigned and RTDetrV2 output differences in v5 are different model regressions." + "reason": "Missing CLIP attentions assignment and RTDetrV2 output drift are distinct regressions." }, { "left": "issue:30064", "right": "issue:36010", "accept": false, - "reason": "Image processor void-segmentation-map failure and the `GenerationMixin` import error are unrelated." + "reason": "Completely unrelated: segmentation-map processing vs a GenerationMixin import error." }, { "left": "issue:43653", "right": "issue:43720", "accept": false, - "reason": "BigBird special-token registration and BitNet packed-weight loading are entirely different issues." + "reason": "BigBird special-token decoding and BitNet accelerate-loading unpacking are unrelated." }, { "left": "issue:43335", "right": "issue:43756", "accept": false, - "reason": "SwitchTransformers sparse-layer creation and Smollm3 RoPE-layer dropping are unrelated model-config bugs." + "reason": "SwitchTransformers sparse-layer creation and Smollm3 RoPE-layer dropping are different architecture bugs." }, { "left": "issue:43716", "right": "issue:44008", "accept": false, - "reason": "Mistral-3 image preprocessor dtype mismatch and Gemma 3n audio-mask name collision are different problems." + "reason": "Mistral-3 dtype mismatch and Gemma 3n variable collision are separate issues." }, { "left": "issue:43295", "right": "issue:45834", "accept": false, - "reason": "A processor/tokenizer regression and a Kosmos2.5 long-OCR index error do not share the same underlying cause." + "reason": "Processor/tokenizer regression and Kosmos2.5 OCR index error are unrelated." }, { "left": "issue:43550", "right": "issue:45072", "accept": false, - "reason": "Bamba torch.compile/SDPA failure and dtype mismatches in SwitchTransformers/TimmWrapperModel are distinct." + "reason": "Bamba torch.compile/SDPA failure and dtype mismatch in other models are different bugs." }, { "left": "issue:43550", "right": "issue:43819", "accept": false, - "reason": "Bamba SDPA/torch.compile issue and DAC latent conversion mismatch are unrelated." + "reason": "Bamba torch.compile/SDPA and DAC latent mismatch are unrelated." }, { "left": "issue:43322", "right": "issue:45685", "accept": false, - "reason": "Llava Next segfault and MPS histogram int support error are not the same bug." + "reason": "Segmentation fault on Llava Next loading and an MPS histogram not implemented error are unrelated runtime issues." } ] }, "evaluator_result": { - "accept": true, - "feedback": "The analyst stays conservative and grounds the summary in the packet. The single accepted pair, issue:42886 and issue:44843, is plausibly the same offline-tokenizer / `_patch_mistral_regex` failure, and the other soft pairs are reasonably rejected as distinct bugs rather than over-merged." + "accept": false, + "feedback": "The cluster summary is mostly grounded, but it is not conservative enough on the Mistral-related subcluster. In particular, `issue:45081` and `issue:45356` are only loosely connected by `fix_mistral_regex`; the packet gives different failure descriptions (a `backend_tokenizer` AttributeError vs. Kimi codec/warning regression) and no shared targets or filenames, so treating them as a clear duplicate pair is overstated. The `42886`/`44843` pair is more plausible, but the overall output still accepts a soft edge that the packet does not clearly support." }, "evaluator_used": true, - "retried": false, + "retried": true, "accepted_nontrivial_soft_edge": true, "error_kind": null, "error_message": null @@ -4741,8 +4726,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -4828,9 +4813,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -5031,8 +5016,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -5120,9 +5105,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -5329,8 +5314,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -5419,9 +5404,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -5634,8 +5619,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -5725,9 +5710,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -5949,8 +5934,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -6037,9 +6022,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -6243,8 +6228,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -6339,9 +6324,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -6590,8 +6575,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -6683,9 +6668,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -6916,8 +6901,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -7006,9 +6991,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -7221,8 +7206,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -7310,9 +7295,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -7528,8 +7513,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -7623,9 +7608,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -7871,8 +7856,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -7961,9 +7946,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -8182,8 +8167,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -8272,9 +8257,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -8487,8 +8472,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -8575,9 +8560,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -8778,8 +8763,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -8867,9 +8852,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -9076,8 +9061,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -9166,9 +9151,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -9387,8 +9372,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -9477,9 +9462,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -9698,8 +9683,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -9789,9 +9774,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -10016,8 +10001,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -10105,9 +10090,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -10314,8 +10299,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -10403,9 +10388,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -10618,8 +10603,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -10708,9 +10693,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 47, @@ -10929,8 +10914,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -11023,9 +11008,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -11349,14 +11334,14 @@ "issue:43278|issue:45137", "issue:44734|issue:45103" ], - "prepared_review_unit_hash": "730eb66fc0268471083c5f17f988d6e7d699dda0f25f9a5955360f320d9e2a11", + "prepared_review_unit_hash": "cac00b0d0602e6f1ddcd2530eeddf288650349207212cbda1a0e0d5a7e350041", "component_budget": { "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -11371,204 +11356,204 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is heavily over-broadened: most items are unrelated issues that only share superficial subsystem or model-family similarity. The only clear duplicate pair is the Qwen3ForCausalLM VRAM leak report.", - "confidence": 0.88, - "canonical_issue_reason": "issue:42673 is the clearest representative of the only true duplicate pair, with a detailed report and higher discussion activity than its twin.", + "summary": "Mostly unrelated issue reports; the only clear duplicate pair is the identical Qwen3ForCausalLM VRAM leak report. No pull requests are present.", + "confidence": 0.94, + "canonical_issue_reason": "issue:42673 is the best canonical target because issue:43704 has the exact same title and describes the same Qwen3ForCausalLM VRAM leak in multiple dataloader threads.", "canonical_pr_reason": null, - "best_issue_reason": "issue:42673 is the strongest global issue candidate because it matches an exact duplicate, has the fuller description, and appears earlier than the matching report.", + "best_issue_reason": "Same concrete bug as its duplicate: identical symptom, model, and multithreaded dataloader context, so 42673 is the strongest representative issue.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:42673", "right": "issue:43704", "accept": true, - "reason": "Exact same title and same underlying bug: Qwen3ForCausalLM VRAM leak in multiple dataloader threads." + "reason": "Exact same bug report and title; clearly the same Qwen3ForCausalLM VRAM leak issue." }, { "left": "issue:45310", "right": "issue:45375", "accept": false, - "reason": "Both involve Qwen3.5 MoE, but one is a from_pretrained error and the other is a missing config field; too different to merge." + "reason": "Both touch Qwen3.5/Moe config loading, but one is a generic from_pretrained error and the other is a specific missing-field/config-schema problem; not enough to call the same bug." }, { "left": "issue:43873", "right": "issue:44945", "accept": false, - "reason": "Different failures: quantization offloading vs incorrect outputs under pipeline parallelism." + "reason": "Different failures: quantization/offloading vs incorrect output under pipeline parallelism." }, { "left": "issue:41628", "right": "issue:44351", "accept": false, - "reason": "Both are import errors, but for different symbols and different code paths." + "reason": "Both are import errors, but for different symbols and different root causes." }, { "left": "issue:42831", "right": "issue:44222", "accept": false, - "reason": "Accuracy regression in FineGrainedFP8 is not the same bug as FP8 save_pretrained failure." + "reason": "Both mention FP8, but one is an accuracy issue and the other is a save_pretrained/MoE persistence bug." }, { "left": "issue:44945", "right": "issue:45753", "accept": false, - "reason": "Pipeline-parallel output corruption and Qwen3_5 image infinite loop are unrelated symptoms." + "reason": "Pipeline-parallel incorrect output and a Qwen3_5 image-specific infinite loop are different concrete problems." }, { "left": "issue:44222", "right": "issue:45310", "accept": false, - "reason": "Saving a quantized MoE model and loading Qwen3.5 MoE with from_pretrained are different concrete problems." + "reason": "FP8 save_pretrained and Qwen3.5 from_pretrained loading are unrelated bugs." }, { "left": "issue:43673", "right": "issue:44188", "accept": false, - "reason": "Generation cache handling during chunked_prefill is unrelated to attention-kernel divergence under torch.compile." + "reason": "Generation cache missing in chunked prefill and attention-kernel divergence under torch.compile are distinct code paths." }, { "left": "issue:36010", "right": "issue:41628", "accept": false, - "reason": "Different missing imports: GenerationMixin versus AutoImageProcessor." + "reason": "Both are import failures, but for different APIs and unrelated breakages." }, { "left": "issue:42994", "right": "issue:44164", "accept": false, - "reason": "Quantized model saving and extra_state save/from_pretrained handling are different serialization bugs." + "reason": "Saving quantized models and handling extra_state in save/from_pretrained are separate issues." }, { "left": "issue:42175", "right": "issue:45712", "accept": false, - "reason": "Packaging/backend dependency issue is not the same as leftover dummy classes leaking into dir(transformers)." + "reason": "Packaging backend dependency issue vs leftover dummy-class hygiene issue; not the same bug." }, { "left": "issue:42222", "right": "issue:45200", "accept": false, - "reason": "VitPose breakage and Gemma 4 mm_token_type_ids behavior are unrelated model-specific issues." + "reason": "VitPose model breakage and Gemma 4 token-type defaults are unrelated." }, { "left": "issue:43232", "right": "issue:44792", "accept": false, - "reason": "Generation kwargs update after sync_gpus is unrelated to a janus image-generation test failure." + "reason": "Generation kwargs update bug and a Janus image-generation test failure do not match." }, { "left": "issue:44038", "right": "issue:45310", "accept": false, - "reason": "Different model families and different load failures; too broad to treat as one bug." + "reason": "Same broad Qwen/Moe area, but different model variants and different failure modes/version regressions." }, { "left": "issue:43377", "right": "issue:44242", "accept": false, - "reason": "Missing padding-mask support in MIMI encoder is unrelated to MoE load-balancing loss computation." + "reason": "MIMI padding-mask batching bug vs MoE load-balancing loss omission are different problems." }, { "left": "issue:43526", "right": "issue:45561", "accept": false, - "reason": "BeitImageProcessorFast label reduction bug is unrelated to the xdist captured_info.txt race." + "reason": "BeitImageProcessorFast label reduction bug vs xdist race on captured_info.txt are unrelated." }, { "left": "issue:43504", "right": "issue:45561", "accept": false, - "reason": "Legacy-field pretrained loading failure is unrelated to the testing-utils race." + "reason": "Legacy-field model loading issue vs pytest-xdist race are different bugs." }, { "left": "issue:44496", "right": "issue:45561", "accept": false, - "reason": "Unrecognized model/config loading error is unrelated to the xdist file race." + "reason": "Unrecognized model/config issue vs testing-utils race condition are unrelated." }, { "left": "issue:44964", "right": "issue:45137", "accept": false, - "reason": "Phi-4 multimodal loading and DeepSpeed ZeRO3 deque error are unrelated." + "reason": "Phi-4 multimodal load failure and DeepSpeed ZeRO3 deque pop error are different code paths." }, { "left": "issue:42175", "right": "issue:44933", "accept": false, - "reason": "Backend dependency installation and nonexistent image_utils import are unrelated." + "reason": "Backend dependency install issue vs missing image_utils import are unrelated." }, { "left": "issue:42898", "right": "issue:44568", "accept": false, - "reason": "Both are tokenization regressions, but they affect different behaviors and models; not the same bug." + "reason": "Tokenizer cleanup behavior change and mDeBERTa special-token addition bug are different regressions." }, { "left": "issue:30990", "right": "issue:35141", "accept": false, - "reason": "Sentence Transformers loading hang and embedding reinitialization in post_init are unrelated." + "reason": "Sentence Transformers loading hang and embedding reinit after resize are unrelated." }, { "left": "issue:43479", "right": "issue:44869", "accept": false, - "reason": "Multimodal config defaults and Whisper timestamp decode crash are unrelated." + "reason": "Multimodal config defaults vs Whisper timestamp decode crash are different issues." }, { "left": "issue:43976", "right": "issue:44933", "accept": false, - "reason": "Python version compatibility and missing import from image_utils are unrelated." + "reason": "Python-version compatibility issue vs missing image_utils import are unrelated." }, { "left": "issue:43479", "right": "issue:44079", "accept": false, - "reason": "Config initialization bug and ModelOutput key assignment bug are unrelated." + "reason": "Config defaulting bug and ModelOutput key assignment bug do not share the same underlying defect." }, { "left": "issue:44188", "right": "issue:44373", "accept": false, - "reason": "Attention-kernel branching under torch.compile is unrelated to a docstring issue." + "reason": "Attention-kernel divergence and a wrong docstring are not the same bug." }, { "left": "issue:43784", "right": "issue:43824", "accept": false, - "reason": "Missing nn import in sentence-transformers and missing Qwen2_5_VLForConditionalGeneration import are different import problems." + "reason": "sentence-transformers import NameError and Qwen2_5_VL import error are unrelated." }, { "left": "issue:44623", "right": "issue:45310", "accept": false, - "reason": "processor.save_pretrained missing files is unrelated to Qwen3.5 MoE from_pretrained failure." + "reason": "Missing processor.save_pretrained files and Qwen3.5 from_pretrained failure are different save/load problems." }, { "left": "issue:42898", "right": "issue:45812", "accept": false, - "reason": "Tokenization behavior change and Granite token-ID regression are different concrete regressions." + "reason": "Tokenizer cleanup behavior change and Granite token-id regression are distinct tokenizer bugs." }, { "left": "issue:43278", "right": "issue:45137", "accept": false, - "reason": "Embedding dtype drift in evaluation is unrelated to the ZeRO3 deque error." + "reason": "BF16-to-FP32 evaluation dtype change and ZeRO3 deque underflow are unrelated." }, { "left": "issue:44734", "right": "issue:45103", "accept": false, - "reason": "KV-cache continuation crash in transformers serve is unrelated to the auto_docstring annotation crash." + "reason": "KV-cache continuation crash in transformers serve and an auto_docstring annotations crash are unrelated." } ] }, "evaluator_result": { "accept": true, - "feedback": "Grounded and conservative. The one accepted pair, issue:42673 vs issue:43704, is supported by identical titles plus the fuller/earlier report on 42673. The remaining soft-edge rejections stay cautious and do not overclaim duplication from superficial similarity." + "feedback": "Grounded in the packet: the only clearly matching pair is the identical Qwen3ForCausalLM VRAM-leak issue titles, and the remaining soft-edge verdicts are conservative. No pull requests are present." }, "evaluator_used": true, "retried": false, @@ -11675,14 +11660,14 @@ "issue:44261|issue:45092", "issue:45092|issue:45357" ], - "prepared_review_unit_hash": "bf533617fb1ce050b8172560b7b6a38b804bd6acfb10c63eff1c7e7c236f00d6", + "prepared_review_unit_hash": "437a461831010131a6112a124645b25b8b32d19487fa9cb43f2a1998b9430527", "component_budget": { "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -11697,222 +11682,198 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "These issues are only loosely related by broad Transformers themes (serialization/save_pretrained, tokenizer regressions, multimodal/model-loading bugs). None of the soft pairs look like the same underlying bug or change, so I rejected all of them. No PRs are present.", - "confidence": 0.94, - "canonical_issue_reason": "Issue 44222 is the strongest concrete bug report in the loose save_pretrained/serialization subcluster, with the clearest actionable failure mode among the candidates.", + "summary": "This cluster is very heterogeneous. The soft links mostly reflect broad topical similarity (tokenizers, save_pretrained, multimodal models, training/deepspeed, Qwen/GLM internals) rather than the same underlying bug or change, so I would not merge any of the candidate pairs.", + "confidence": 0.93, + "canonical_issue_reason": "No strong canonical duplicate hub stands out; the items span unrelated bugs. If forced to anchor the cluster, issue:44222 is the most concrete self-contained bug report, but it does not subsume the rest.", "canonical_pr_reason": null, - "best_issue_reason": "Issue 44222 is the best overall anchor here: it is a specific, bug-shaped report rather than a question or broad symptom, even though the surrounding items are not true duplicates.", + "best_issue_reason": "issue:44222 is the most actionable standalone report in the set, with a clear symptom and scope, but it is not a true umbrella for the cluster.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:30333", "right": "issue:33290", "accept": false, - "reason": "MLflow failure reporting is unrelated to the deepspeed/Adafactor OOM." + "reason": "Both are training-related, but one is an MLflow reporting issue and the other is an OOM bug in Adafactor/DeepSpeed." }, { "left": "issue:43723", "right": "issue:44568", "accept": false, - "reason": "Different tokenizer regressions: load failure vs add_special_tokens/BOS-EOS behavior." + "reason": "Both involve tokenizer behavior in v5, but one is a load/import failure and the other is missing BOS/EOS insertion." }, { "left": "issue:43874", "right": "issue:45092", "accept": false, - "reason": "Missing image-patch API and remote-code meta-init incompatibility are different bugs." + "reason": "Both touch multimodal vision paths, but one is a missing image-patch API and the other is meta-init compatibility for old checkpoints." }, { "left": "issue:42898", "right": "issue:45356", "accept": false, - "reason": "v5 tokenization-space behavior change and Kimi codec handling/warning are distinct tokenizer issues." + "reason": "Both are tokenizer regressions, but the affected code paths and symptoms are different." }, { "left": "issue:44060", "right": "issue:45125", "accept": false, - "reason": "Incorrect tied-weights warning is not the same as a missing _tp_plan for tensor parallelism." + "reason": "Same general Qwen/MoE area, but one is a tied-weights warning and the other is a missing tensor-parallel plan." }, { "left": "issue:42898", "right": "issue:44779", "accept": false, - "reason": "Both involve tokenizers, but one is a v5 spacing change and the other is DeepSeek output corruption." + "reason": "Both concern tokenizers in v5, but they describe different models and different failure modes." }, { "left": "issue:45405", "right": "issue:45712", "accept": false, - "reason": "A released-version dependency bump and leftover dummy classes are unrelated maintenance problems." + "reason": "Completely unrelated: a version pin issue versus leftover dummy class cleanup." }, { "left": "issue:42898", "right": "issue:44448", "accept": false, - "reason": "Both are version regressions, but they affect different models and symptoms." + "reason": "Both mention v5 output changes, but one is a tokenizer-space behavior change and the other is a Pegasus model output regression." }, { "left": "issue:44295", "right": "issue:45656", "accept": false, - "reason": "Buffer-reading error and double optimizer-step behavior are unrelated code paths." + "reason": "Unrelated: a buffer/position_ids bug versus a DeepSpeed double optimizer-step issue." }, { "left": "issue:44933", "right": "issue:45712", "accept": false, - "reason": "Missing image_utils import and dummy_pt_objects leftovers are different import/export issues." + "reason": "Different areas entirely: missing image_utils import versus dummy class leakage." }, { "left": "issue:44222", "right": "issue:44297", "accept": false, - "reason": "Both mention save_pretrained, but one is FP8 MoE saving and the other is tokenizer metadata mismatch." + "reason": "Both involve save_pretrained, but one is model FP8 serialization and the other is tokenizer metadata." }, { "left": "issue:44297", "right": "issue:45656", "accept": false, - "reason": "Tokenizer save metadata and duplicate optimizer steps are unrelated." + "reason": "No shared code path; tokenizer save metadata and optimizer stepping are unrelated." }, { "left": "issue:43116", "right": "issue:44190", "accept": false, - "reason": "Example-script multi-label inference failure and local dataset loading failure are different problems." + "reason": "Both are example-script issues, but one is classification output logic and the other is local dataset loading." }, { "left": "issue:43479", "right": "issue:43644", "accept": false, - "reason": "Config initialization with None and junk-filled non-persistent buffers are different bugs." + "reason": "Different bugs: multimodal config initialization versus non-persistent buffer initialization." }, { "left": "issue:44222", "right": "issue:44623", "accept": false, - "reason": "Both are serialization-related, but model save_pretrained FP8 behavior is not the same as processor.save_pretrained missing files." + "reason": "Both mention save_pretrained, but they concern different object types and different missing behavior." }, { "left": "issue:44206", "right": "issue:45103", "accept": false, - "reason": "Unsupported feature argument in a feature extractor is unrelated to auto_docstring annotation crashes." + "reason": "These are unrelated crashes: a feature-extractor argument mismatch versus auto-docstring annotation handling." }, { "left": "issue:42831", "right": "issue:44623", "accept": false, - "reason": "FineGrainedFP8 accuracy problems are unrelated to missing processor save files." + "reason": "Accuracy drift in FineGrainedFP8 is not the same problem as processor save_pretrained missing files." }, { "left": "issue:41628", "right": "issue:42757", "accept": false, - "reason": "Different missing imports from different packages/modules." + "reason": "Both are import errors, but they come from different packages and different missing symbols." }, { "left": "issue:22355", "right": "issue:28282", "accept": false, - "reason": "transformers.onnx import failure and missing PyTorch are separate import-environment issues." + "reason": "Different import failures with different root causes: missing transformers.onnx module versus missing PyTorch dependency." }, { "left": "issue:30333", "right": "issue:36331", "accept": false, - "reason": "MLflow not reported on failure is unrelated to the CustomTrainer compute_loss signature breakage." + "reason": "Both involve training, but one is MLflow reporting and the other is a Trainer API signature mismatch." }, { "left": "issue:44297", "right": "issue:45310", "accept": false, - "reason": "Tokenizer save metadata mismatch and Qwen3.5 from_pretrained failure are different load/save bugs." + "reason": "Tokenizer save metadata and Qwen3.5 from_pretrained failures are different code paths and symptoms." }, { "left": "issue:30333", "right": "issue:34689", "accept": false, - "reason": "Training-job reporting and Llama 3.2 Vision model loading regression are unrelated." + "reason": "Unrelated: MLflow/job-state reporting versus a model-loading regression for Llama 3.2 Vision." }, { "left": "issue:43232", "right": "issue:45834", "accept": false, - "reason": "Generation kwargs update after sync_gpus is not the same as an OCR long-input index error." + "reason": "Different generation/ocr bugs with no shared underlying mechanism." }, { "left": "issue:44222", "right": "issue:44484", "accept": false, - "reason": "A concrete save_pretrained bug and a question about max_shard_size are not duplicates." - }, - { - "left": "issue:43723", - "right": "issue:44977", - "accept": false, - "reason": "Tokenizer loading in v5 and Qwen3.5 flash-attention generation failure are different issues." - }, - { - "left": "issue:28282", - "right": "issue:30333", - "accept": false, - "reason": "Missing PyTorch during AutoModel import and MLflow failure reporting are unrelated." - }, - { - "left": "issue:44263", - "right": "issue:45092", - "accept": false, - "reason": "torch.split return handling and InternVL2 meta-init incompatibility are different model bugs." - }, - { - "left": "issue:30333", - "right": "issue:35141", - "accept": false, - "reason": "MLflow reporting on failure is unrelated to embedding reinitialization in post_init." - }, - { - "left": "issue:44483", - "right": "issue:45406", - "accept": false, - "reason": "Chat-completions request rejection and Gemma4Processor missing _tokenizer are different API/processor failures." + "reason": "Save_pretrained FP8 serialization is unrelated to max_shard_size configuration semantics." }, { "left": "issue:44263", "right": "issue:44360", "accept": false, - "reason": "Same DSA indexer area, but one is a split-return bug and the other is a missing ReLU concern." + "reason": "Same DSA indexer area, but the reported defects are different: torch.split handling versus missing ReLU." }, { "left": "issue:44373", "right": "issue:44485", "accept": false, - "reason": "A wrong docstring and a RoPE implementation discussion are not the same bug." + "reason": "A docstring issue is not the same as a RoPE implementation bug." }, { "left": "issue:44295", "right": "issue:44623", "accept": false, - "reason": "Reading a registered buffer and processor save_pretrained file loss are unrelated." + "reason": "Different concerns: buffer registration/reads versus processor serialization file output." }, { "left": "issue:44514", "right": "issue:45103", "accept": false, - "reason": "Chat-template batching crash and auto_docstring annotation crash are different code paths." + "reason": "Unrelated crashes: batched chat template padding versus auto-docstring annotation introspection." }, { "left": "issue:44261", "right": "issue:45092", "accept": false, - "reason": "MLA layernorm epsilon precision mismatch and InternVL2 meta-init incompatibility are different bugs." + "reason": "Different multimodal bugs: RMS norm epsilon precision versus old checkpoint meta-init incompatibility." + }, + { + "left": "issue:44263", + "right": "issue:44360", + "accept": false, + "reason": "Related only by the DSA indexer subsystem; the code-path problem is not the same." }, { "left": "issue:45092", "right": "issue:45357", "accept": false, - "reason": "Old InternVL2 meta-init issues and Qwen3.5 visual-key save_pretrained regressions are unrelated." + "reason": "Both relate to Qwen/vision loading, but one is meta-init compatibility and the other is save_pretrained key serialization." } ] }, @@ -12019,14 +11980,14 @@ "issue:39401|issue:44488", "issue:44261|issue:44373" ], - "prepared_review_unit_hash": "465037515db1d7074a19085e82715cb2cbb6e677b68bee01aac2d202febe3f65", + "prepared_review_unit_hash": "30b87966ccfe3bacad7e3e73a020438f1bc6bd50a5d87403e7158e22de089491", "component_budget": { "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -12041,210 +12002,204 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is highly heterogeneous: most items are unrelated tokenizer/import/config/regression reports with only superficial similarity, so there is no strong duplicate core. The broadest representative is the v5 tokenizer-loading regression issue.", - "confidence": 0.41, - "canonical_issue_reason": "Issue 43723 is the broadest, most central tokenizer-loading regression in the set and overlaps the most with other tokenizer-related reports, though the cluster is still quite mixed.", + "summary": "These issues are mostly separate tokenizer, import, config, and model-behavior regressions that only overlap at a broad subsystem level. None of the soft pairs look like the same underlying bug or change, so all soft edges are rejected.", + "confidence": 0.79, + "canonical_issue_reason": "issue:43723 is the broadest and most central tokenizer-loading regression in this set, so it is the best umbrella issue even though the others are distinct bugs.", "canonical_pr_reason": null, - "best_issue_reason": "Issue 43723 is the best single representative because it is broad, user-facing, and closest to the common tokenizer-loading theme seen in several nearby candidates.", + "best_issue_reason": "issue:43723 has the clearest general scope around AutoTokenizer loading failures in v5 and the strongest fit as a representative issue for the cluster.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43723", "right": "issue:44871", "accept": false, - "reason": "Both are tokenizer-related, but one is a generic v5 loading failure and the other is a Gemma-3 EOS config mismatch; different bugs." + "reason": "Both are tokenizer-related, but one is a general loading regression and the other is an eos_token_id configuration mismatch; not the same bug." }, { "left": "issue:30990", "right": "issue:38175", "accept": false, - "reason": "Completely different symptoms and code paths: Sentence Transformers loading hang vs zero probabilities in a SigLIP2 model." + "reason": "Unrelated symptoms: Sentence Transformers loading hang versus zero probabilities in a SigLIP model." }, { "left": "issue:43723", "right": "issue:44991", "accept": false, - "reason": "Both mention tokenizer loading in v5, but they are model-specific failures and do not look like the same underlying bug." + "reason": "Both mention tokenizer loading, but they concern different models and likely different failure modes." }, { "left": "issue:43723", "right": "issue:44442", "accept": false, - "reason": "Different tokenizer implementations and failure modes; no evidence they share the same concrete bug." + "reason": "General AutoTokenizer load issue versus a specific FastSpeech2ConformerTokenizer failure; not the same underlying defect." }, { "left": "issue:41628", "right": "issue:43824", "accept": false, - "reason": "Both are import errors, but for different missing symbols and different subsystems." + "reason": "Both are import errors, but for different symbols and model classes; too broad to treat as duplicates." }, { "left": "issue:43479", "right": "issue:44913", "accept": false, - "reason": "Unrelated config bugs: multimodal default config initialization vs GPTNeoX rotary_pct persistence." + "reason": "Different config bugs: default sub-config initialization versus rotary_pct not persisting on reload." }, { "left": "issue:42831", "right": "issue:45310", "accept": false, - "reason": "Accuracy issue in FineGrainedFP8 vs a Qwen3.5 MoE from_pretrained error; no shared code-path." + "reason": "Accuracy regression in FineGrainedFP8 and a from_pretrained error for Qwen3.5 Moe are unrelated." }, { "left": "issue:44263", "right": "issue:44373", "accept": false, - "reason": "One is a tensor-split indexing bug, the other is a docstring issue; not the same change or bug." + "reason": "A tensor-splitting bug and a wrong docstring are not the same issue." }, { "left": "issue:39401", "right": "issue:44568", "accept": false, - "reason": "Both are tokenizer-related, but one is offset mapping in Qwen3 and the other is missing BOS/EOS insertion for mdeberta-v3." + "reason": "Both involve tokenization behavior, but one is offset mapping and the other is add_special_tokens/BOS-EOS handling." }, { "left": "issue:42831", "right": "issue:44038", "accept": false, - "reason": "Different problems: numeric accuracy with FP8 vs Qwen3-VL-Moe loading/compatibility." + "reason": "FineGrainedFP8 accuracy issue and Qwen3-VL-Moe loading/bug report are different problems." }, { "left": "issue:42757", "right": "issue:45278", "accept": false, - "reason": "One specific missing import from huggingface_hub and one broad import-error report; too general to treat as the same bug." + "reason": "A missing import from huggingface_hub and a broad set of import errors are only superficially related." }, { "left": "issue:42886", "right": "issue:42994", "accept": false, - "reason": "Offline tokenizer cache loading vs quantized model saving failure are unrelated." + "reason": "Offline tokenizer cache loading and quantized model saving are distinct code paths." }, { "left": "issue:43232", "right": "issue:45593", "accept": false, - "reason": "Generation state update bug vs D-FINE auxiliary-loss behavior; different areas entirely." + "reason": "Generation kwargs handling and D-FINE auxiliary-loss behavior are unrelated." }, { "left": "issue:43756", "right": "issue:44060", "accept": false, - "reason": "Different model-specific architecture/config issues: RoPE layer handling vs tied-weights warning." + "reason": "Different Qwen/Smollm model bugs with no clear shared code path." }, { "left": "issue:43576", "right": "issue:43784", "accept": false, - "reason": "Broken env CLI command vs import-time NameError in sentence-transformers; not the same bug." + "reason": "A broken CLI command and a NameError during sentence-transformers import are unrelated." }, { "left": "issue:42831", "right": "issue:45753", "accept": false, - "reason": "FineGrainedFP8 accuracy issue and Qwen3_5 infinite loop are unrelated." + "reason": "FP8 accuracy regression and an infinite loop on a specific image are different issues." }, { "left": "issue:42886", "right": "issue:43066", "accept": false, - "reason": "Both involve tokenizer/v5 behavior, but one is offline cache loading and the other is tokenizer decoder typing; different failures." + "reason": "Offline cache loading and wrong tokenizer decoder type are tokenizer-related but not the same bug." }, { "left": "issue:43122", "right": "issue:44521", "accept": false, - "reason": "Both involve tokenization/chat templates, but one is version-dependent tokenization drift and the other is multimodal assistant mask generation." + "reason": "Both touch tokenization, but one is a version-to-version tokenization change and the other is assistant_mask generation for multimodal chat templates." }, { "left": "issue:42371", "right": "issue:44263", "accept": false, - "reason": "TF32 settings warning vs a torch.split indexing issue; unrelated." + "reason": "TF32 settings guidance and a torch.split return-value bug are unrelated." }, { "left": "issue:22355", "right": "issue:28218", "accept": false, - "reason": "No module import error for transformers.onnx vs extra-space tokenizer behavior; unrelated." + "reason": "A missing transformers.onnx module and an added-token spacing bug are unrelated." }, { "left": "issue:44117", "right": "issue:44704", "accept": false, - "reason": "Both are from_pretrained-related, but one is TOKENIZER_MAPPING_NAMES handling and the other is AutoProcessor kwargs forwarding." + "reason": "Tokenizer mapping handling and AutoProcessor kwargs forwarding are different loader bugs." }, { "left": "issue:30990", "right": "issue:33357", "accept": false, - "reason": "Loading hang in sentence-transformers vs MacOS bus error in a CLIP model; not the same bug." + "reason": "A loading hang and a MacOS bus error are distinct failures." }, { "left": "issue:43232", "right": "issue:44743", "accept": false, - "reason": "Both touch generation/cache behavior, but the concrete failures differ: sync_gpus kwargs update vs recurrent-state reset in modular_qwen3_5." + "reason": "Generation state update after sync_gpus and recurrent-state resetting in qwen3 modular code are not the same defect." }, { "left": "issue:44295", "right": "issue:44297", "accept": false, - "reason": "Reading a bufferized position_ids tensor vs tokenizer_class mismatch in saved config; unrelated." + "reason": "Reading position_ids as a buffer and tokenizer_class mismatch on save are unrelated tokenizer/model issues." }, { "left": "issue:44704", "right": "issue:45290", "accept": false, - "reason": "AutoProcessor kwargs propagation vs chat_template crash with tool-call assistant messages; different code paths." + "reason": "AutoProcessor kwargs forwarding and apply_chat_template crashing on tool-call messages are different code paths." }, { "left": "issue:42907", "right": "issue:44945", "accept": false, - "reason": "Dequantized model saving failure vs incorrect output under pipeline parallelism; unrelated." + "reason": "Quantized model saving failure and incorrect output under pipeline parallelism are unrelated." }, { "left": "issue:41628", "right": "issue:45278", "accept": false, - "reason": "Specific missing AutoImageProcessor import vs broad import-error report; not the same concrete issue." + "reason": "A specific AutoImageProcessor import error and broad post-upgrade import errors are not the same bug." }, { "left": "issue:43122", "right": "issue:44514", "accept": false, - "reason": "Both are chat/tokenization issues, but one is version tokenization drift and the other is a batched apply_chat_template crash." + "reason": "Version-dependent tokenization differences and batched apply_chat_template crashes are different issues." }, { "left": "issue:43976", "right": "issue:44945", "accept": false, - "reason": "Python version compatibility problem vs pipeline-parallelism output bug; unrelated." + "reason": "Python version compatibility and pipeline-parallelism output correctness are unrelated." }, { "left": "issue:43784", "right": "issue:45600", "accept": false, - "reason": "Import NameError in sentence-transformers vs CI break from removed auto_mappings config references; different failures." + "reason": "Import-time NameError in sentence-transformers and removed auto_mappings references are different problems." }, { "left": "issue:39401", "right": "issue:44488", "accept": false, - "reason": "Both are tokenizer loading-related, but the affected models and failure modes are different." + "reason": "A Qwen3 offset-mapping bug and a failure to load a specific model are not the same issue." }, { "left": "issue:44261", "right": "issue:44373", "accept": false, - "reason": "Precision/config issue in MLA q_a_layernorm vs a docstring correction; not duplicates." - }, - { - "left": "issue:44514", - "right": "issue:44521", - "accept": false, - "reason": "Both involve apply_chat_template, but one is a batched-padding crash and the other is zero assistant masks for multimodal inputs; distinct bugs." + "reason": "A precision bug in MLA q_a_layernorm and a wrong docstring for position_ids are unrelated." } ] }, @@ -12263,8 +12218,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -12356,9 +12311,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -12589,8 +12544,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -12680,9 +12635,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -12992,14 +12947,14 @@ "issue:38617|issue:42757", "issue:44261|issue:44485" ], - "prepared_review_unit_hash": "038bff0d50e6d97c538201a87f7830d2f1570620b2515b4ef0e337d42638bd91", + "prepared_review_unit_hash": "51fbfce399d3084a546659130372e6d2a6c5b5020106574d5280cd7334dc9428", "component_budget": { "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -13014,222 +12969,222 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is noisy and mostly contains unrelated issue reports that only look similar at the title level. I found no soft pair that clearly describes the same underlying bug/change, so I would not merge any of them.", - "confidence": 0.57, - "canonical_issue_reason": "Issue 45071 is the broadest, most central Transformers-v5 regression in the set and best represents the cluster\u2019s general compatibility-breakage theme.", + "summary": "This cluster is a loose mix of Transformers regressions and import/model-loading bugs that only look similar at the title level. The soft edges mostly share generic v5 migration or backend wording, but they describe different code paths and failure modes, so none should be merged as duplicates.", + "confidence": 0.64, + "canonical_issue_reason": "If a single anchor is needed, issue 45372 is the most active and current report, but the cluster is too heterogeneous to have a true canonical duplicate target.", "canonical_pr_reason": null, - "best_issue_reason": "Issue 45071 is the most reusable representative: it is a cross-cutting v5 regression, whereas most others are narrower model- or feature-specific failures.", + "best_issue_reason": "Issue 45372 is the best single issue to represent the cluster only because it is open, concrete, and has follow-up traffic; it is not a strong duplicate hub for the other reports.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43937", "right": "issue:44496", "accept": false, - "reason": "GLM generation-config validation vs model-type/config loading failure; different bugs." + "reason": "Both are v5-era validation/loading complaints, but one is a GenerationConfig error for GLM-5 and the other is an unrecognized model/config problem for OLMo; different failures." }, { "left": "issue:42175", "right": "issue:42222", "accept": false, - "reason": "Backend installation packaging issue vs VitPose model breakage; unrelated." + "reason": "These are unrelated backend/model issues: TensorFlow not being included with a torch install versus broken vitpose model files." }, { "left": "issue:44483", "right": "issue:45464", "accept": false, - "reason": "Both mention chat/completions, but one is a generic request rejection and the other is streaming inference on a specific model." + "reason": "Both mention chat/completions, but one is a general API request rejection and the other is a streaming inference failure on a specific Qwen model." }, { "left": "issue:38617", "right": "issue:44220", "accept": false, - "reason": "ImportError for a config symbol vs audio feature extraction issue; no concrete overlap." + "reason": "These are different import/audio bugs: a missing transformers symbol versus a feature extraction helper issue." }, { "left": "issue:42371", "right": "issue:42831", "accept": false, - "reason": "TF32 settings warning is unrelated to FineGrainedFP8 accuracy." + "reason": "TF32 configuration guidance and FineGrainedFP8 accuracy are different numeric-behavior problems, not the same bug." }, { "left": "issue:43066", "right": "issue:43122", "accept": false, - "reason": "Tokenizer decoder-type regression vs tokenization output change; related area, not same bug." + "reason": "Related to tokenizer behavior in v5, but one is a wrong decoder type and the other is a tokenization change regression; not the same underlying defect." }, { "left": "issue:44075", "right": "issue:45656", "accept": false, - "reason": "SGD argument handling vs duplicated optimizer step under deepspeed; different failure modes." + "reason": "Both involve optimizers stepping incorrectly, but the concrete issue differs: unused SGD args versus deepspeed calling step twice." }, { "left": "issue:43673", "right": "issue:45341", "accept": false, - "reason": "Generation cache regression vs a testing-utils bug; unrelated." + "reason": "Generation cache availability in v5 is unrelated to a small testing_utils bug." }, { "left": "issue:43012", "right": "issue:43577", "accept": false, - "reason": "Precision warning during compile vs dtype not propagating on load; same theme, different code paths." + "reason": "Both concern precision/dtype handling, but one is a PyTorch warning during bf16 compilation and the other is BLIP2 loading leaving submodules in float32." }, { "left": "issue:43874", "right": "issue:44261", "accept": false, - "reason": "Missing image-patch helper vs MLA layernorm epsilon precision issue; unrelated." + "reason": "Both touch numerical precision, but one is a missing image-patch helper in GLM46V and the other is an MLA layernorm epsilon mismatch." }, { "left": "issue:44484", "right": "issue:45478", "accept": false, - "reason": "Save/sharding question vs Qwen3.5 MoE from_pretrained failure; not the same bug." + "reason": "Both are Qwen3.5-related failures, but one asks about save_pretrained shard sizing while the other is a from_pretrained error in a MoE model." }, { "left": "issue:44410", "right": "issue:45478", "accept": false, - "reason": "Missing projections in one model family vs a different Qwen3.5 MoE loading error." + "reason": "Layer-0 missing projections in qwen3next is a model-architecture bug, not the same as Qwen3.5 MoE loading failure." }, { "left": "issue:43097", "right": "issue:44493", "accept": false, - "reason": "Removed tie-embeddings API vs unexpected position-id key warnings; different regressions." + "reason": "These are both v5 migration regressions, but one is removal of an encoder-decoder tie flag and the other is unexpected position-id keys during loading." }, { "left": "issue:44263", "right": "issue:44485", "accept": false, - "reason": "torch.split return behavior vs RoPE implementation; too different." + "reason": "Different GLM-family bugs: torch.split return handling versus RoPE implementation concerns." }, { "left": "issue:41628", "right": "issue:45071", "accept": false, - "reason": "AutoImageProcessor import failure vs PretrainedConfig type-checking regression; different API breaks." + "reason": "Both are API/regression reports, but one is a missing AutoImageProcessor import and the other is a PreTrainedConfig type-checking break." }, { "left": "issue:43012", "right": "issue:43408", "accept": false, - "reason": "bfloat16 precision warning vs SAM3 model-type mismatch warning; unrelated." + "reason": "Both mention warnings, but one is bf16 compile precision and the other is a model-type mismatch warning for sam3_video/sam3_tracker." }, { "left": "issue:42222", "right": "issue:45459", "accept": false, - "reason": "VitPose model breakage vs tokenizer error masking without protobuf; no shared bug." + "reason": "A broken vitpose model and a tokenizer error-hiding import_protobuf_decode_error problem are unrelated." }, { "left": "issue:43441", "right": "issue:44829", "accept": false, - "reason": "Both involve flash attention, but different models and symptoms; not mergeable as one fix." + "reason": "Ministral FlashAttention failure and flash_attention_3 training degeneration are both attention-related, but they are different model/code paths and symptoms." }, { "left": "issue:45071", "right": "issue:45542", "accept": false, - "reason": "Config type checking regression vs TensorBoard-only backend issue; unrelated." + "reason": "PreTrainedConfig type checking and a missing TensorFlow backend caused by tensorboard-only installs are not the same regression." }, { "left": "issue:43931", "right": "issue:44410", "accept": false, - "reason": "Qwen3-VL weight-shape mismatch vs qwen3next missing projections; different concrete problems." + "reason": "These are separate model-loading shape mismatch and qwen3next projection bugs." }, { "left": "issue:43335", "right": "issue:45071", "accept": false, - "reason": "SwitchTransformers sparse-layer construction bug vs generic config type-checking regression." + "reason": "SwitchTransformers sparse-layer construction and PreTrainedConfig type checking are distinct config/constructor issues." }, { "left": "issue:43931", "right": "issue:44038", "accept": false, - "reason": "Different Qwen3-VL model-loading failures; no evidence of the same root cause." + "reason": "Qwen3-VL weight-shape mismatch and Qwen3-VL-Moe from_pretrained failures are different model variants and failure modes." }, { "left": "issue:45003", "right": "issue:45341", "accept": false, - "reason": "Unsafe sys.modules access vs testing-utils bug; unrelated." + "reason": "Unsafe sys.modules access in modeling_utils is unrelated to the minor testing_utils bug." }, { "left": "issue:43097", "right": "issue:43452", "accept": false, - "reason": "Embedding-tie deprecation vs gguf_file tokenizer/model loading breakage; different issues." + "reason": "A removed tie_embeddings flag and gguf_file breaking tokenizer/model loading are separate regression areas." }, { "left": "issue:43986", "right": "issue:45341", "accept": false, - "reason": "AutoProcessor/torchvision crash vs testing-utils bug; unrelated." + "reason": "A missing torchvision dependency path in AutoProcessor is unrelated to a small testing utility bug." }, { "left": "issue:44117", "right": "issue:44987", "accept": false, - "reason": "Tokenizer mapping None vs a specific model load failure; too little evidence of the same root cause." + "reason": "TOKENIZER_MAPPING_NAMES returning None and loading physical-intelligence/fast are different tokenizer versus model-loading failures." }, { "left": "issue:43986", "right": "issue:44484", "accept": false, - "reason": "Video AutoProcessor crash vs max_shard_size question; unrelated." + "reason": "AutoProcessor crashing without torchvision and max_shard_size behavior are unrelated." }, { "left": "issue:43408", "right": "issue:44368", "accept": false, - "reason": "Model-type warning for sam3_tracker vs embedding-tie warning in Qwen3.5 fine-tuning; different configs." + "reason": "Model-type mismatch warnings and tie_word_embeddings warnings are separate configuration warnings." }, { "left": "issue:43874", "right": "issue:44263", "accept": false, - "reason": "Image patch helper missing vs torch.split return values; unrelated." + "reason": "Both are GLM-related, but one is a missing image-patch helper and the other is a torch.split return-value issue." }, { "left": "issue:43976", "right": "issue:45071", "accept": false, - "reason": "Python version support regression vs config type-checking regression; different breakages." + "reason": "Python version support in 5.1.0 and PreTrainedConfig type checking are unrelated release regressions." }, { "left": "issue:43986", "right": "issue:45372", "accept": false, - "reason": "Torchvision-dependent processor crash vs missing ReasoningEffort import; unrelated." + "reason": "AutoProcessor without torchvision and the ReasoningEffort import error affect different processor/dependency paths." }, { "left": "issue:44829", "right": "issue:45245", "accept": false, - "reason": "Flash-attention training degeneration vs category cardinality limit; unrelated." + "reason": "FlashAttention-3 training degeneration and the 2^24 category limit are completely different problems." }, { "left": "issue:43531", "right": "issue:45341", "accept": false, - "reason": "Qwen3-MoE sliding-window bug vs testing-utils bug; no overlap." + "reason": "Qwen3-MoE sliding_window behavior is unrelated to the testing_utils bug." }, { "left": "issue:38617", "right": "issue:42757", "accept": false, - "reason": "Different import errors from different modules; not the same failure." + "reason": "Both are import errors, but they involve different missing symbols from different packages and code paths." }, { "left": "issue:44261", "right": "issue:44485", "accept": false, - "reason": "LayerNorm epsilon precision issue vs RoPE implementation discussion; distinct bugs." + "reason": "MLA q_a_layernorm epsilon precision and GLM-5 RoPE implementation are separate numerical/model-architecture issues." } ] }, @@ -13248,8 +13203,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -13344,9 +13299,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -13595,8 +13550,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -13689,9 +13644,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -14005,14 +13960,14 @@ "issue:43502|issue:44222", "issue:42831|issue:43874" ], - "prepared_review_unit_hash": "53f14646eae41d8dcb897d8b34f53dca965f70842fb83ee84170c1736180031c", + "prepared_review_unit_hash": "a421655fb8e437156fc30156ba9116792704a22b9552940f22730118d06d0655", "component_budget": { "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -14027,204 +13982,210 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "These are mostly unrelated issue reports, with a few thematically similar loading/regression bugs but no true duplicate-level matches among the soft pairs.", - "confidence": 0.66, - "canonical_issue_reason": "issue:43502 is the clearest, most actionable bug report in the set: it describes a concrete, user-visible regression with an obvious repro and fix target.", + "summary": "This cluster is mostly a loose set of Transformers issues sharing broad loading/version-regression language, but the concrete bugs are different enough that none of the soft pairs should be merged as duplicates.", + "confidence": 0.53, + "canonical_issue_reason": "Issue 45850 is the broadest open umbrella here: it describes widespread breakage across many models after 5.0, making it the best representative for a mixed regression cluster.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43502 is the best issue anchor because it is specific, reproducible, and directly actionable rather than broad or model-specific.", + "best_issue_reason": "Issue 45850 is the best global issue candidate because it is open, generic, and can plausibly collect related regression reports without being tied to one model-specific failure.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43867", "right": "issue:45357", "accept": false, - "reason": "Different failures: state_dict ordering during model loading vs incorrect visual encoder keys on save_pretrained." + "reason": "Both involve model loading/saving regressions, but one is about sorted state_dict loading and the other about incorrect visual encoder keys in save_pretrained." }, { "left": "issue:43010", "right": "issue:45850", "accept": false, - "reason": "One is a specific no_grad decoration request; the other is a broad complaint that many models break." + "reason": "Decorating cache/layer update with no_grad is a specific API fix, not the broad multi-model regression described in 45850." }, { "left": "issue:43992", "right": "issue:45478", "accept": false, - "reason": "Both are loading problems, but they target different models and different root causes." + "reason": "Both mention from_pretrained failures, but they target different Qwen/UMT5 model paths and likely different missing weights." }, { "left": "issue:43010", "right": "issue:43976", "accept": false, - "reason": "Unrelated: cache update decoration vs a Python-version compatibility failure." + "reason": "One is a no_grad behavior bug; the other is a Python-version compatibility failure." }, { "left": "issue:44938", "right": "issue:45381", "accept": false, - "reason": "Different areas: Python 3.14 import/load failure vs Qwen2.5-VL video position-id behavior." + "reason": "Python 3.14 import/load failure and Qwen2.5-VL vision_position_ids corruption are unrelated code paths." }, { "left": "issue:43618", "right": "issue:44077", "accept": false, - "reason": "Different model/API issues: CLIPOutput attentions vs patchtsmixer post_init policy." + "reason": "CLIPOutput attentions assignment and PatchTSMixer post_init validation are unrelated model API issues." }, { "left": "issue:42490", "right": "issue:44075", "accept": false, - "reason": "Local model path warning and SGD optimizer arguments are unrelated bugs." + "reason": "A local-path import warning and SGD optimizer args being ignored are unrelated." }, { "left": "issue:36296", "right": "issue:43575", "accept": false, - "reason": "Both mention tensor parallelism, but one is a generic training bug and the other is a specific OOM on loading a Qwen model." + "reason": "Both mention tensor parallelism, but one is a training bug and the other is an OOM when loading a specific Qwen2 checkpoint." }, { "left": "issue:44360", "right": "issue:45440", "accept": false, - "reason": "Different code paths: DSA indexer activation issue vs DeepseekV3MoE implementation divergence." + "reason": "DSA indexer ReLU discussion and DeepSeekV3 implementation divergence are different components and problems." }, { "left": "issue:28282", "right": "issue:33453", "accept": false, - "reason": "ImportError for missing PyTorch and tokenizer loading regression are not the same defect." + "reason": "PyTorch missing ImportError and tokenizer loading regression are separate failure modes." }, { "left": "issue:43867", "right": "issue:44484", "accept": false, - "reason": "Model load failure from sorted state_dict is unrelated to the save_pretrained shard-size question." + "reason": "Both touch save/load APIs, but one is a loading bug and the other is a max_shard_size behavior question." }, { "left": "issue:43618", "right": "issue:44857", "accept": false, - "reason": "CLIPOutput attentions assignment and AMP/CUDA loss crash are different bugs." + "reason": "CLIPOutput attentions and float16 AMP crash in LwDetrImageLoss are unrelated." }, { "left": "issue:43618", "right": "issue:44977", "accept": false, - "reason": "Different subsystems and symptoms: CLIPOutput field regression vs Qwen3.5 flash-attention generation bug." + "reason": "A missing attentions field is not the same as Qwen3.5 flash-attention generation corruption." }, { "left": "issue:42890", "right": "issue:44964", "accept": false, - "reason": "A flaky test due to missing set_seed is unrelated to a model-loading failure." + "reason": "A flaky test due to missing set_seed is unrelated to a Phi-4 multimodal model loading regression." }, { "left": "issue:42890", "right": "issue:44112", "accept": false, - "reason": "Both are CI/test-related, but the failure modes and affected components differ." + "reason": "Both are test-related, but they fail for different reasons and on different model suites." }, { "left": "issue:42890", "right": "issue:45600", "accept": false, - "reason": "Test seeding issue and auto_mappings stale config references are separate problems." + "reason": "A seed-sensitive integration test is unrelated to auto_mappings referencing removed Sam3LiteText configs." }, { "left": "issue:43502", "right": "issue:44062", "accept": false, - "reason": "Local_files_only network leakage and AddedToken keyword collision are unrelated." + "reason": "Local_files_only network leakage and AddedToken keyword collision are different tokenizer/runtime bugs." }, { "left": "issue:43723", "right": "issue:44556", "accept": false, - "reason": "Both involve version changes and loading, but one is tokenizer loading and the other is checkpoint reload after upgrade." + "reason": "Both are loading regressions, but one is tokenizer-specific and the other is checkpoint reload after version upgrade." }, { "left": "issue:42890", "right": "issue:45341", "accept": false, - "reason": "A flaky model integration test is not the same as a generic testing_utils bug." + "reason": "A flaky model test and a testing_utils bug are not the same underlying issue." }, { "left": "issue:44222", "right": "issue:45125", "accept": false, - "reason": "FP8 save_pretrained behavior and missing tensor-parallel plan are different feature gaps." + "reason": "FP8 save_pretrained and missing _tp_plan for tensor parallelism affect different code paths." }, { "left": "issue:44387", "right": "issue:44556", "accept": false, - "reason": "Int4 memory regression and checkpoint reload incompatibility are different issues." + "reason": "CUDA memory growth under int4 quantization is unrelated to checkpoint reload breakage after upgrade." }, { "left": "issue:42548", "right": "issue:44351", "accept": false, - "reason": "Both are import errors, but for different symbols and likely different export fixes." + "reason": "Both are import errors, but they concern different missing symbols and likely different export regressions." }, { "left": "issue:43010", "right": "issue:43065", "accept": false, - "reason": "No_grad on update methods and a dummy Conv2d in Sam3PixelDecoder are unrelated." + "reason": "A no_grad decoration bug and a dummy Conv2d in Sam3PixelDecoder are unrelated." }, { "left": "issue:43502", "right": "issue:44485", "accept": false, - "reason": "local_files_only network requests and GLM-5 RoPE implementation are unrelated defects." + "reason": "Network access despite local_files_only and GLM-5 RoPE implementation are unrelated." }, { "left": "issue:44355", "right": "issue:44749", "accept": false, - "reason": "Compiled-Python-file errors and post-upgrade inference slowdown are different problems." + "reason": "Compiled Python file errors and slow filtering after upgrade are unrelated." }, { "left": "issue:45125", "right": "issue:45478", "accept": false, - "reason": "Both are Qwen3.5 MoE-related, but one is missing tensor-parallel metadata and the other is a generic from_pretrained failure." + "reason": "They are both Qwen3.5 MoE loading issues, but the first is specifically missing _tp_plan while the second is a broader from_pretrained error without enough evidence of the same root cause." }, { "left": "issue:43010", "right": "issue:43232", "accept": false, - "reason": "Generation state update timing and no_grad decoration are different code paths." + "reason": "Cache/layer no_grad decoration and generation kwargs sync behavior are different generation internals." }, { "left": "issue:30990", "right": "issue:31795", "accept": false, - "reason": "Stuck model loading and documentation confusion are not the same bug." + "reason": "Sentence Transformers loading hang and forward-doc confusion are unrelated." }, { "left": "issue:45125", "right": "issue:45375", "accept": false, - "reason": "Qwen3.5 tensor-parallel metadata and missing vision config field are separate configuration bugs." + "reason": "Missing _tp_plan for Qwen3.5 MoE and missing deepstack_visual_indexes in Qwen3_5MoeVisionConfig are different configuration problems." }, { "left": "issue:30333", "right": "issue:33453", "accept": false, - "reason": "MLFlow reporting on failed jobs and tokenizer loading regression are unrelated." + "reason": "MLflow reporting on failed training and tokenizer loading regression are unrelated." }, { "left": "issue:43299", "right": "issue:44556", "accept": false, - "reason": "Both mention upgrade-related loading regressions, but they affect different models and failure modes." + "reason": "Both are version-related loading failures, but they affect different model families and likely different loading code paths." }, { "left": "issue:43502", "right": "issue:44222", "accept": false, - "reason": "Network access despite local_files_only and FP8 save_pretrained behavior are not duplicates." + "reason": "A local-files-only network bug and FP8 save_pretrained issues are unrelated." + }, + { + "left": "issue:42831", + "right": "issue:43874", + "accept": false, + "reason": "FineGrainedFP8 accuracy and GLM46V image-processor method absence are different model-specific bugs." } ] }, @@ -14243,8 +14204,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -14343,9 +14304,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -14606,8 +14567,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -14709,9 +14670,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 48, @@ -15005,8 +14966,8 @@ "review_unit_count": 47, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-10000-400", "nodes": [ @@ -15052,9 +15013,9 @@ "node_count": 400, "item_count": 400, "soft_pair_count": 1499, - "serialized_chars": 772968, - "estimated_input_tokens": 193242, - "estimated_eval_tokens": 386740 + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 }, "budget": { "node_count": 19, @@ -15183,8 +15144,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-13244-3", "nodes": [ @@ -15254,8 +15215,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-26350-2", "nodes": [ @@ -15317,8 +15278,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-27088-4", "nodes": [ @@ -15417,8 +15378,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-36817-15", "nodes": [ @@ -15731,8 +15692,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-37934-6", "nodes": [ @@ -15847,8 +15808,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-41084-2", "nodes": [ @@ -15910,8 +15871,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-43598-2", "nodes": [ @@ -15973,8 +15934,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-43684-2", "nodes": [ @@ -16036,8 +15997,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-43835-4", "nodes": [ @@ -16139,8 +16100,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-44371-5", "nodes": [ @@ -16261,8 +16222,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-45145-2", "nodes": [ @@ -16381,8 +16342,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-45672-2", "nodes": [ @@ -16444,8 +16405,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-45762-2", "nodes": [ @@ -16507,8 +16468,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-45764-2", "nodes": [ @@ -16570,8 +16531,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "llm", - "cache_hit": false, + "source": "cache", + "cache_hit": true, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-45869-5", "nodes": [ @@ -16741,178 +16702,178 @@ "pull_request:44827|pull_request:45922", "pull_request:43251|pull_request:43254" ], - "prepared_review_unit_hash": "a33177a03aea6443bb15300d6f1cb27231bb9e303e6daebbc911e61023395bd2", + "prepared_review_unit_hash": "b55d20d144ef80c3c13b8376a5e390e335227a9a88d87f7488b18563914f97af", "component_budget": { "node_count": 16, "item_count": 16, "soft_pair_count": 29, - "serialized_chars": 46699, - "estimated_input_tokens": 11675, - "estimated_eval_tokens": 23606 + "serialized_chars": 46702, + "estimated_input_tokens": 11676, + "estimated_eval_tokens": 23608 }, "budget": { "node_count": 16, "item_count": 16, "soft_pair_count": 29, - "serialized_chars": 46699, - "estimated_input_tokens": 11675, - "estimated_eval_tokens": 23606 + "serialized_chars": 46702, + "estimated_input_tokens": 11676, + "estimated_eval_tokens": 23608 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "The cluster is mostly heterogeneous: model additions, docs updates, refactors, and unrelated fixes. Only one true duplicate pair stands out: the two `fixed_cross_entropy` PRs that both implement the same kwargs support in `src/transformers/loss/loss_utils.py` for issue 43240.", - "confidence": 0.97, + "summary": "This cluster is mostly a grab bag of unrelated PRs (new model additions, docs, and refactors). The only true duplicate pair is the two `fixed_cross_entropy` changes; the rest should stay separate.", + "confidence": 0.93, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 43251 is the clearest canonical representative of the only real duplicate pair: it targets issue 43240 and makes the concrete `fixed_cross_entropy` kwargs change in `loss_utils.py`.", + "canonical_pr_reason": "PR 43251 is the best canonical PR for the only real duplicate pair: it directly targets issue 43240 and contains the fuller `fixed_cross_entropy` kwargs-forwarding change, while 43254 is a narrower variant.", "best_issue_reason": null, - "best_pr_reason": "PR 43251 is the strongest single PR to keep as representative because it directly fixes the linked loss issue with the exact same code-path change as 43254.", + "best_pr_reason": "PR 43251 is the strongest representative of the actual duplicate change because it is the more complete implementation of the same cross-entropy fix and has the explicit issue linkage; the other PRs are not duplicates of it.", "soft_edge_verdicts": [ { "left": "pull_request:39895", "right": "pull_request:43424", "accept": false, - "reason": "Unrelated changes: VideoPrism model/docs addition vs ExecuTorch dynamic-shape export test." + "reason": "Videoprism model addition vs ExecuTorch dynamic-shape test; different subsystems and no shared bug/change." }, { "left": "pull_request:41855", "right": "pull_request:43823", "accept": false, - "reason": "Different artifacts: Mistral tokenizer method additions vs adding a new MobileLLM model." + "reason": "Mistral tokenizer API additions vs adding a new MobileLLM model; unrelated changes." }, { "left": "pull_request:41855", "right": "pull_request:43424", "accept": false, - "reason": "Tokenizer API additions vs ExecuTorch export test; no shared bug or change." + "reason": "Tokenizer utility fix vs export test coverage; not the same code-path problem." }, { "left": "pull_request:42668", "right": "pull_request:43823", "accept": false, - "reason": "Processor robustness fix vs new model implementation; not the same underlying change." + "reason": "Processor auto-mapping robustness work vs new MobileLLM model implementation; different underlying changes." }, { "left": "pull_request:43251", "right": "pull_request:43823", "accept": false, - "reason": "Loss utility kwargs fix vs MobileLLM model add; unrelated." + "reason": "Cross-entropy helper fix vs new model addition; no overlap beyond being Transformers PRs." }, { "left": "pull_request:43251", "right": "pull_request:44827", "accept": false, - "reason": "Both touch training-related code, but cross-entropy kwargs support and Mistral4 test fixes are different changes." + "reason": "Loss-function kwargs fix and Mistral4 test/model work are unrelated." }, { "left": "pull_request:43424", "right": "pull_request:45919", "accept": false, - "reason": "ExecuTorch export test and Sapiens2 model addition are unrelated." + "reason": "ExecuTorch export testing vs adding Sapiens2; different goals and code paths." }, { "left": "pull_request:43424", "right": "pull_request:45899", "accept": false, - "reason": "ExecuTorch exportability test vs docs-only continuous batching update; different code paths." + "reason": "Export dynamic-shape test vs docs-only continuous batching note; not a duplicate change." }, { "left": "pull_request:43424", "right": "pull_request:44827", "accept": false, - "reason": "Dynamic-shape export test vs Mistral4 test/model work; no shared fix." + "reason": "ExecuTorch export test and Mistral4 test/model fixes are unrelated." }, { "left": "pull_request:43424", "right": "pull_request:43823", "accept": false, - "reason": "ExecuTorch export testing and new model scaffolding are unrelated." + "reason": "Export testing vs MobileLLM model addition; not the same underlying bug/change." }, { "left": "pull_request:43823", "right": "pull_request:45655", "accept": false, - "reason": "New MobileLLM model vs config-resolution ordering fix; not the same bug." + "reason": "New MobileLLM model vs config-resolution refactor; unrelated." }, { "left": "pull_request:43823", "right": "pull_request:45922", "accept": false, - "reason": "New model addition vs lru-cache memory leak fix in vision models; different changes." + "reason": "New MobileLLM model vs vision-model lru cache leak fix; different code paths." }, { "left": "pull_request:43823", "right": "pull_request:43995", "accept": false, - "reason": "New model addition vs Falcon output-collection refactor; unrelated." + "reason": "MobileLLM model addition vs Falcon output-interface refactor; no duplicate relationship." }, { "left": "pull_request:43995", "right": "pull_request:45213", "accept": false, - "reason": "Falcon refactor vs model-creation skill docs; not the same artifact or fix." + "reason": "Falcon model refactor vs model-creation skill/docs; completely different changes." }, { "left": "pull_request:43995", "right": "pull_request:44827", "accept": false, - "reason": "Falcon interface refactor and Mistral4 test fix are different code-path changes." + "reason": "Falcon output-collection refactor vs Mistral4 test fixes; not the same fix." }, { "left": "pull_request:43995", "right": "pull_request:45899", "accept": false, - "reason": "Falcon model refactor vs docs-only decode fast-path note; unrelated." + "reason": "Falcon model refactor vs continuous-batching docs update; unrelated." }, { "left": "pull_request:44827", "right": "pull_request:45899", "accept": false, - "reason": "Mistral4 test/model work vs docs-only continuous batching update; no common bug." + "reason": "Mistral4 test/model fixes vs docs for decode fast path; not the same underlying change." }, { "left": "pull_request:44827", "right": "pull_request:45170", "accept": false, - "reason": "Mistral4 test fix vs CLIP-family layernorm typo mapping; different changes." + "reason": "Mistral4 test fixes vs `layrnorm` renaming pass; different models and problems." }, { "left": "pull_request:44827", "right": "pull_request:45795", "accept": false, - "reason": "Mistral4 test/model changes vs docs for audio/video processors; unrelated." + "reason": "Mistral4 test fixes vs docs for audio/video processors; unrelated." }, { "left": "pull_request:44827", "right": "pull_request:45655", "accept": false, - "reason": "Mistral4 work vs config resolution order fix; different underlying issue." + "reason": "Mistral4 fix vs config-resolution order fix; different issues." }, { "left": "pull_request:45170", "right": "pull_request:45899", "accept": false, - "reason": "Layernorm typo rename vs docs-only batching fast-path description; unrelated." + "reason": "CLIP-like weight renaming vs docs-only batching writeup; no shared code path." }, { "left": "pull_request:44827", "right": "pull_request:45213", "accept": false, - "reason": "Mistral4 tests/model code vs model-creation skill docs; not the same change." + "reason": "Shared auto/modeling_auto filename is incidental; Mistral4 tests and model-creation skill are unrelated." }, { "left": "pull_request:43823", "right": "pull_request:45213", "accept": false, - "reason": "New MobileLLM model vs model-creation skill docs; no shared code-path problem." + "reason": "New model addition vs skill/doc scaffolding for model creation; not the same change." }, { "left": "pull_request:45213", "right": "pull_request:45899", "accept": false, - "reason": "Skill docs for model creation vs continuous batching docs update; unrelated." + "reason": "Model-creation skill PR vs continuous-batching docs; unrelated documentation areas." }, { "left": "pull_request:45795", @@ -16924,31 +16885,31 @@ "left": "pull_request:43823", "right": "pull_request:45795", "accept": false, - "reason": "New MobileLLM model vs docs for processor components; unrelated." + "reason": "MobileLLM model addition vs processor docs; no duplicate underlying bug/change." }, { "left": "pull_request:43823", "right": "pull_request:45899", "accept": false, - "reason": "New MobileLLM model vs decode fast-path docs; no common underlying change." + "reason": "New model addition vs continuous-batching docs update; unrelated." }, { "left": "pull_request:44827", "right": "pull_request:45922", "accept": false, - "reason": "Mistral4 test/model work vs vision-model cache-leak fix; unrelated." + "reason": "Mistral4 test fixes vs vision-model memory-leak fix; different model families and defects." }, { "left": "pull_request:43251", "right": "pull_request:43254", "accept": true, - "reason": "Same file, same function, same issue 43240, and both add the same `cross_entropy` kwargs support; these are duplicate fixes for the same code path." + "reason": "Both PRs patch `fixed_cross_entropy` in the same file for the same issue (43240) and differ only in how the supported kwargs are threaded through; they could be merged into one fix." } ] }, "evaluator_result": { "accept": true, - "feedback": "Grounded and conservative overall. The only accepted duplicate pair is well supported by the packet: both PRs modify `src/transformers/loss/loss_utils.py`, target issue 43240, and implement the same `fixed_cross_entropy` kwargs expansion. The other soft-edge pairs are rejected in a cautious, file- and change-based way without overclaiming a shared bug." + "feedback": "Grounded and conservative overall. The only pair that looks like a real duplicate bugfix is PR 43251/43254: both edit `src/transformers/loss/loss_utils.py` for issue 43240 and are mergeable into one `fixed_cross_entropy` kwargs-forwarding fix. The other pair verdicts are appropriately negative given the very different subsystems and file sets, even where the similarity heuristic was noisy." }, "evaluator_used": true, "retried": false, @@ -16975,12 +16936,12 @@ "soft_pairs": [ "pull_request:40546|pull_request:42781" ], - "prepared_review_unit_hash": "cc55c2bca5fcae75723397ff187f7da679e99a1eb725387e982aee32259b7fcb", + "prepared_review_unit_hash": "56d44b2db28df2867e1839c908b5c8418f06309ef086e934b11c8aa32c1edb0b", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 7225, + "serialized_chars": 7226, "estimated_input_tokens": 1807, "estimated_eval_tokens": 3870 }, @@ -16988,7 +16949,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 7225, + "serialized_chars": 7226, "estimated_input_tokens": 1807, "estimated_eval_tokens": 3870 }, @@ -16997,18 +16958,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are related VibeVoice PRs, but they are not duplicates: PR 40546 is the base VibeVoice implementation, while PR 42781 adds a separate realtime variant with acoustic-tokenizer support and additional generation/feature-extraction work.", - "confidence": 0.89, + "summary": "Two related but distinct VibeVoice PRs: one implements the base VibeVoice integration, the other adds VibeVoice Realtime with extra acoustic-tokenizer and generation/model changes. They overlap on shared package wiring, but they are not the same underlying change.", + "confidence": 0.91, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 40546 is the better canonical PR because it is the broader, foundational VibeVoice implementation and already covers the core model integration, docs, and Auto-class wiring.", + "canonical_pr_reason": "pull_request:42781 is the more specific and more expansive VibeVoice follow-up, covering the realtime path plus the acoustic tokenizer and additional generation/model work; pull_request:40546 is the earlier base VibeVoice implementation.", "best_issue_reason": null, - "best_pr_reason": "PR 40546 is the best representative because it establishes the main VibeVoice codepath; PR 42781 is a later draft focused on realtime-specific extensions rather than the same concrete change.", + "best_pr_reason": "pull_request:42781 is the better representative of the current VibeVoice work because it covers the broader, newer realtime feature set and includes the shared VibeVoice integration pieces.", "soft_edge_verdicts": [ { "left": "pull_request:40546", "right": "pull_request:42781", "accept": false, - "reason": "They overlap on VibeVoice plumbing, but 42781 is a distinct realtime follow-up with extra acoustic-tokenizer and generation changes. It is not the same underlying fix/change and would not be safely merged as one PR." + "reason": "Reject: both PRs share VibeVoice infrastructure, but 42781 adds a distinct realtime variant and acoustic-tokenizer/model changes, so they are not the same concrete change and would not realistically be merged as one PR." } ] }, @@ -17038,7 +16999,7 @@ "soft_pairs": [ "pull_request:40755|pull_request:41224" ], - "prepared_review_unit_hash": "da5861b5337a1a214beebad06de6747884ec825e5e5e829d4975ef0e399c0a51", + "prepared_review_unit_hash": "69a18bf0dde5addc97f4a6e448a32d2e77172290805e51d6bb21137618e88373", "component_budget": { "node_count": 2, "item_count": 2, @@ -17060,18 +17021,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These two pull requests are unrelated feature additions: one adds covariate-based forecasting to TimesFM, the other adds DINOv3 image classification support. They do not appear to fix the same code-path problem or be mergeable into one PR.", - "confidence": 0.98, + "summary": "The two PRs are unrelated feature additions in different model families and code paths, so this is not a duplicate cluster.", + "confidence": 0.99, "canonical_issue_reason": null, - "canonical_pr_reason": null, + "canonical_pr_reason": "No canonical PR: #40755 adds TimesFM covariate forecasting, while #41224 adds DINOv3 image classification support; they are distinct features in different subsystems.", "best_issue_reason": null, - "best_pr_reason": null, + "best_pr_reason": "Neither PR is a good global representative for the cluster because they address unrelated enhancements and should not be merged or deduplicated together.", "soft_edge_verdicts": [ { "left": "pull_request:40755", "right": "pull_request:41224", "accept": false, - "reason": "Different models, different subsystems, and different goals: TimesFM covariate forecasting vs. DINOv3 image classification. No shared underlying bug/change." + "reason": "Different models and different changes: TimesFM forecasting with covariates vs DINOv3 image classification support. They do not fix the same underlying bug or implement the same change." } ] }, @@ -17166,46 +17127,46 @@ "pull_request:41797|pull_request:45186", "pull_request:45186|pull_request:45630" ], - "prepared_review_unit_hash": "6629ed434b5547738597a9e45f3ae5335e5c39796b1266be3a0ae38a01bff6f7", + "prepared_review_unit_hash": "06859588dc075a9e835878ff034261da2c07225f583000bb155ab791287ef33b", "component_budget": { "node_count": 3, "item_count": 3, "soft_pair_count": 2, - "serialized_chars": 10408, - "estimated_input_tokens": 2602, - "estimated_eval_tokens": 5460 + "serialized_chars": 10409, + "estimated_input_tokens": 2603, + "estimated_eval_tokens": 5462 }, "budget": { "node_count": 3, "item_count": 3, "soft_pair_count": 2, - "serialized_chars": 10408, - "estimated_input_tokens": 2602, - "estimated_eval_tokens": 5460 + "serialized_chars": 10409, + "estimated_input_tokens": 2603, + "estimated_eval_tokens": 5462 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are three independent model-addition PRs (DeepSeek OCR, Isaac, and Kimi2-6). They share only the standard Transformers model-onboarding scaffolding, not the same underlying change.", - "confidence": 0.97, + "summary": "These are three separate new-model integration PRs with overlapping Transformers boilerplate, not the same underlying change. The shared files are generic registration/documentation plumbing, while each PR targets a different model family and code path.", + "confidence": 0.98, "canonical_issue_reason": null, - "canonical_pr_reason": null, + "canonical_pr_reason": "PR 45186 is the most central representative of the cluster because it connects to both other PRs and reflects the shared model-integration scaffolding.", "best_issue_reason": null, - "best_pr_reason": null, + "best_pr_reason": "PR 45186 is the best single representative here for triage because it sits at the center of the similarity graph, even though the PRs are not duplicates.", "soft_edge_verdicts": [ { "left": "pull_request:41797", "right": "pull_request:45186", "accept": false, - "reason": "Different model integrations: DeepSeek OCR vs Isaac. Shared files are generic registry/docs/template updates typical of new model PRs, not the same code-path fix or mergeable duplicate change." + "reason": "Different model additions: DeepSeek OCR vs Isaac. The overlap is only in generic docs/auto-registration files, not the same concrete bug or feature." }, { "left": "pull_request:45186", "right": "pull_request:45630", "accept": false, - "reason": "Different model integrations: Isaac vs Kimi2-6. The overlap is broad onboarding machinery and docs, but the concrete model code, configs, and tests are for unrelated models, so these are not duplicates." + "reason": "Different model additions: Isaac vs Kimi2-6. Shared filenames are boilerplate integration points, but the PRs implement separate model code paths." } ] }, @@ -17235,12 +17196,12 @@ "soft_pairs": [ "pull_request:42975|pull_request:42976" ], - "prepared_review_unit_hash": "b9ddd2a1be960ca3b58f11ff5f827b7d5486751a47a888b378a036a7264e13a4", + "prepared_review_unit_hash": "4e069303b72d97d1faa2436a437e3bea501c536af27852685da957224ed1cc06", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 5934, + "serialized_chars": 5935, "estimated_input_tokens": 1484, "estimated_eval_tokens": 3224 }, @@ -17248,7 +17209,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 5934, + "serialized_chars": 5935, "estimated_input_tokens": 1484, "estimated_eval_tokens": 3224 }, @@ -17257,18 +17218,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are related only at the level of GitHub Actions maintenance, but they target different action families and different changes: PR 42975 upgrades core GitHub actions for Node 24 compatibility, while PR 42976 updates docker/build-push-action to v6 across Docker image workflows. They should not be merged as duplicates.", + "summary": "These PRs both update GitHub Actions dependencies, but they address different underlying changes: one is a Node 24 compatibility sweep for core workflow actions, while the other is a set of version bumps focused on docker/build-push-action and related workflow updates. They are related in theme, not duplicates.", "confidence": 0.96, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 42975 is the clearer canonical representative: it has the more specific stated goal (Node 24 compatibility) and a broader, more coherent set of workflow action upgrades.", + "canonical_pr_reason": "PR #42975 is the better representative of the cluster theme because it targets the core GitHub Actions runtime compatibility problem across many workflows (checkout/cache/upload-artifact).", "best_issue_reason": null, - "best_pr_reason": "PR 42975 is the best overall representative of the cluster because its scope is more self-contained and explicitly framed, whereas PR 42976 is a separate Docker action version bump.", + "best_pr_reason": "PR #42975 is the stronger representative because it has the broader, more foundational workflow-action compatibility scope; PR #42976 is narrower and mainly upgrades Docker build actions.", "soft_edge_verdicts": [ { "left": "pull_request:42975", "right": "pull_request:42976", "accept": false, - "reason": "Different underlying changes: one upgrades core GitHub Actions for Node 24 compatibility, the other bumps docker/build-push-action to v6 in Docker image workflows. Same general maintenance area, but not the same fix/change." + "reason": "Do not merge: they upgrade different actions and fix different concerns. #42975 is about Node 24 compatibility for core actions, while #42976 bumps docker/build-push-action and a few workflow versions; they are not the same concrete change." } ] }, @@ -17311,12 +17272,12 @@ "pull_request:44601|pull_request:45838", "pull_request:45490|pull_request:45838" ], - "prepared_review_unit_hash": "443824e7b861755b7db7690480fdf0858e946a73fac77320496b56b73d50fbbf", + "prepared_review_unit_hash": "1546ffa79f2dd77710fef19d466c113faab47a9633023d3e0b599dc0b51dcf05", "component_budget": { "node_count": 7, "item_count": 7, "soft_pair_count": 9, - "serialized_chars": 23541, + "serialized_chars": 23542, "estimated_input_tokens": 5886, "estimated_eval_tokens": 12028 }, @@ -17324,7 +17285,7 @@ "node_count": 7, "item_count": 7, "soft_pair_count": 9, - "serialized_chars": 23541, + "serialized_chars": 23542, "estimated_input_tokens": 5886, "estimated_eval_tokens": 12028 }, @@ -17333,66 +17294,66 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "This cluster is a loose topical grouping of unrelated PRs: several independent model-addition PRs (ViT NEPA, Molmo, Molmo2, Lfm2 Audio, CTSM, PP-OCRv6 support) plus one distributed pipeline-parallel infrastructure PR. The apparent overlap is mostly boilerplate docs/auto-registry changes, not duplicate code changes.", + "summary": "This cluster is heterogeneous: it mixes several unrelated model-addition PRs plus one distributed-loading feature PR. Despite superficial filename overlap in docs/auto registries, they do not fix the same underlying code-path problem, so no duplicates should be merged.", "confidence": 0.98, "canonical_issue_reason": null, "canonical_pr_reason": null, "best_issue_reason": null, - "best_pr_reason": "PR 43451 is the strongest representative item here: it is a large, highly reviewed model-addition PR with substantial implementation/test coverage. It is not a duplicate of the others, but it is the most developed single PR in the cluster.", + "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "pull_request:42978", "right": "pull_request:43448", "accept": false, - "reason": "Different model families and code paths: ViT NEPA vs Molmo. Shared docs/auto-registration edits are standard boilerplate for model adds, not the same change." + "reason": "Different model families and different implementation areas: ViT NEPA vs Molmo. Shared doc/auto registry edits are just boilerplate for new model support, not the same change." }, { "left": "pull_request:42978", "right": "pull_request:43451", "accept": false, - "reason": "ViT NEPA and Molmo2 are separate model additions with different implementations, processors, and tests; no shared bug or identical feature change." + "reason": "ViT NEPA and Molmo2 are separate model integrations with different configs, modeling code, and tests; not the same bug or feature." }, { "left": "pull_request:42978", "right": "pull_request:45490", "accept": false, - "reason": "ViT NEPA vs CTSM are unrelated model additions. Similar registry/docs edits do not indicate the same underlying change." + "reason": "ViT NEPA is unrelated to CTSM; both are model additions but for different architectures and code paths." }, { "left": "pull_request:43448", "right": "pull_request:43451", "accept": false, - "reason": "Molmo and Molmo2 are related by name but are distinct models with different files and functionality; this is not one fix/feature split across two PRs." + "reason": "Molmo and Molmo2 are distinct model releases with separate packages and tests. Similar naming does not make them the same underlying change." }, { "left": "pull_request:43448", "right": "pull_request:45490", "accept": false, - "reason": "Molmo and CTSM are different model stacks and different changes entirely." + "reason": "Molmo and CTSM are unrelated model additions touching different model code, so they are not duplicates." }, { "left": "pull_request:43451", "right": "pull_request:45490", "accept": false, - "reason": "Molmo2 and CTSM are separate model additions with no shared concrete code-path problem." + "reason": "Molmo2 and CTSM are separate model integrations with no shared concrete bug or code-path fix." }, { "left": "pull_request:43973", "right": "pull_request:45838", "accept": false, - "reason": "Lfm2 Audio adds an audio model, while PP-OCRv6 adds text-recognition models. Different modalities, code paths, and goals." + "reason": "Lfm2 Audio is an audio model addition, while PP-OCRv6 support is text recognition. Different domains, files, and code paths." }, { "left": "pull_request:44601", "right": "pull_request:45838", "accept": false, - "reason": "Pipeline-parallel loading support is core distributed infrastructure; PP-OCRv6 is a model-family addition. They are not the same bug or change." + "reason": "Pipeline-parallel loading support is a distributed systems/runtime change; PP-OCRv6 is a model support PR. They do not address the same problem." }, { "left": "pull_request:45490", "right": "pull_request:45838", "accept": false, - "reason": "CTSM and PP-OCRv6 are different model-family additions with distinct implementations and tests, not a duplicate pair." + "reason": "CTSM model support and PP-OCRv6 text recognition support are distinct model additions with no shared underlying change." } ] }, @@ -17425,7 +17386,7 @@ "pull_request:43291|pull_request:44815", "pull_request:43665|pull_request:44815" ], - "prepared_review_unit_hash": "98ae86de9abe81f55837c73daa31d2407544121d849be89f7be1c0b8f1f86cdd", + "prepared_review_unit_hash": "cfc2ce156458e9974119231631ef6442e0b74100a93b7e5dad56c60b6f210e0a", "component_budget": { "node_count": 3, "item_count": 3, @@ -17447,30 +17408,30 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These three PRs are not duplicates. 43291 is a Whisper tokenizer/test fix, 43665 is mostly test-only noise, and 44815 is a substantive FP8/model-loading dequant change. The only apparent similarity is broad \u201cfix\u201d/test activity, which is not enough to merge them.", - "confidence": 0.96, + "summary": "These PRs are not duplicates: one fixes Whisper decoding/tests, one is a Mistral/FP8 dequantization fix, and one appears to be unrelated test edits in CLIP/ViT. The only plausible representative is the FP8/dequant fix PR.", + "confidence": 0.95, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 44815 is the clearest representative of the cluster because it contains the only substantial code-path fix; the other two PRs are unrelated test/Whisper changes.", + "canonical_pr_reason": "PR 44815 is the most substantive and self-contained bug fix, with a clear code-path focus in loading/FP8 quantization and supporting model/test updates.", "best_issue_reason": null, - "best_pr_reason": "PR 44815 is the strongest standalone PR: it fixes core model-loading / fine-grained FP8 behavior and includes targeted related changes, unlike the other two which are unrelated or test-only.", + "best_pr_reason": "PR 44815 best represents the cluster because it addresses a concrete runtime/path issue across the dequantization and loading flow, unlike the other PRs which are narrower or appear unrelated.", "soft_edge_verdicts": [ { "left": "pull_request:43291", "right": "pull_request:43665", "accept": false, - "reason": "Different areas and different kinds of changes: Whisper tokenizer logic vs CLIP/ViT test edits. No shared bug or mergeable combined fix." + "reason": "Different areas and changes: Whisper tokenizer decoding/tests versus CLIP/ViT test edits. No shared concrete bug or mergeable single fix." }, { "left": "pull_request:43291", "right": "pull_request:44815", "accept": false, - "reason": "Whisper timestamp decoding/tokenization is unrelated to FP8 dequant/model loading. Same generic \u201cfix\u201d label only." + "reason": "Whisper decode behavior is unrelated to Mistral FP8/dequant loading logic; they touch different code paths and would not be one combined fix." }, { "left": "pull_request:43665", "right": "pull_request:44815", "accept": false, - "reason": "43665 is test-only noise for CLIP/ViT, while 44815 fixes FP8 dequant/model loading. Not the same underlying change." + "reason": "CLIP/ViT test-only edits are unrelated to the FP8 dequantization and model-loading changes in 44815, so these are not the same underlying change." } ] }, @@ -17500,7 +17461,7 @@ "soft_pairs": [ "pull_request:43488|pull_request:43532" ], - "prepared_review_unit_hash": "38c12097a507280dfb1541a78d110ddad94ad595ef053af4c08c70103f26b0c1", + "prepared_review_unit_hash": "b6fc18938084cf7282db2e5b613fab1104e182c751d08d9cdb8e7118278e154e", "component_budget": { "node_count": 2, "item_count": 2, @@ -17522,18 +17483,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These two open PRs are not duplicates: they target different changes, and the shared Vit test filename looks incidental. PR 43488 is a small setup/version/test-bot check, while PR 43532 changes CI workflows and testing utilities. Do not merge them together.", - "confidence": 0.97, + "summary": "Two open PRs are grouped only by weak surface similarity, but they change different code paths: one is a repo-bot/versioning test PR, the other rewires CI/comment-diff workflow plumbing. They are not duplicates.", + "confidence": 0.98, "canonical_issue_reason": null, "canonical_pr_reason": null, "best_issue_reason": null, - "best_pr_reason": "If one PR must be singled out as the cluster representative, 43532 is broader and has more context, but this is not a true duplicate cluster.", + "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "pull_request:43488", "right": "pull_request:43532", "accept": false, - "reason": "Different underlying changes: 43488 is about setup.py/dependency/version and a Vit test tweak, while 43532 modifies CI workflows, testing utilities, and related test files. Shared test file alone is not enough to treat them as the same fix." + "reason": "Different underlying changes and goals; shared vit test file is incidental, not evidence of the same fix or mergeable duplicate PRs." } ] }, @@ -17563,12 +17524,12 @@ "soft_pairs": [ "pull_request:43612|pull_request:43613" ], - "prepared_review_unit_hash": "625cd7798fc97138e880b6a7f06ec9f37d583258da591fcc9c2610aeec11d751", + "prepared_review_unit_hash": "9f8a6b2a2229ff17b1dfa226126422bdbc437821fa6472b56eef90d30bf3b29e", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 6082, + "serialized_chars": 6083, "estimated_input_tokens": 1521, "estimated_eval_tokens": 3298 }, @@ -17576,7 +17537,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 6082, + "serialized_chars": 6083, "estimated_input_tokens": 1521, "estimated_eval_tokens": 3298 }, @@ -17585,18 +17546,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are two separate pull requests adding different segmentation pipelines: one for Promptable Concept Segmentation and one for Promptable Visual Segmentation. They share some documentation and registration scaffolding, but the concrete pipeline behavior, model coverage, and test targets differ, so they should not be treated as duplicates.", - "confidence": 0.96, + "summary": "The two pull requests are related feature additions in the same segmentation area, but they implement different pipelines and different code paths: concept segmentation vs visual segmentation. They should not be treated as duplicates.", + "confidence": 0.94, "canonical_issue_reason": null, - "canonical_pr_reason": null, + "canonical_pr_reason": "PR 43613 is the broader, more integrated feature addition, with more model integrations and test coverage, so it is the better representative PR for this cluster.", "best_issue_reason": null, - "best_pr_reason": null, + "best_pr_reason": "PR 43613 is the stronger representative because it covers the larger promptable visual segmentation pipeline surface and includes more supporting files and model hookups.", "soft_edge_verdicts": [ { "left": "pull_request:43612", "right": "pull_request:43613", "accept": false, - "reason": "Different underlying changes: one introduces Promptable Concept Segmentation, the other Promptable Visual Segmentation. Shared docs/auto-registration files are incidental and do not indicate the same code-path fix or a PR that could be merged as one." + "reason": "Different features and code paths: promptable concept segmentation vs promptable visual segmentation. Shared docs/infrastructure do not make them the same underlying change." } ] }, @@ -17626,7 +17587,7 @@ "soft_pairs": [ "pull_request:43651|pull_request:43663" ], - "prepared_review_unit_hash": "60e265cd84b773917a5e3e13cb74aa8d8f067bfbf56046e82da82442e4ba2ef5", + "prepared_review_unit_hash": "f42780f43d885ef13db20945118edd2617314d2499dc3da181cf691e9bbfc30f", "component_budget": { "node_count": 2, "item_count": 2, @@ -17648,18 +17609,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Two open PRs in the same trainer file, but they fix different trainer extension points: one changes gradient-accumulation loss scaling, the other extracts signature-column filtering for overrideability. They are related in subsystem, not duplicates.", - "confidence": 0.96, + "summary": "These are two different trainer.py changes: one addresses gradient-accumulation loss scaling, the other adds an overridable signature-column helper for data filtering. They share a file and subsystem, but not the same underlying bug or mergeable change set.", + "confidence": 0.97, "canonical_issue_reason": null, - "canonical_pr_reason": "No canonical PR in this cluster: the PRs target different code paths and different customization hooks, so they should not be collapsed into one change.", + "canonical_pr_reason": null, "best_issue_reason": null, - "best_pr_reason": "PR #43651 is the more representative concrete bug-fix style change in this cluster because it directly alters training-step loss normalization behavior; however, it still is not a duplicate of #43663.", + "best_pr_reason": "43651 is the more self-contained, user-facing trainer behavior change, but it is not a duplicate of 43663.", "soft_edge_verdicts": [ { "left": "pull_request:43651", "right": "pull_request:43663", "accept": false, - "reason": "Different underlying problems: #43651 adds control over gradient-accumulation loss scaling, while #43663 refactors signature-column inspection for custom trainers/PEFT. Same file and similar theme, but not the same fix." + "reason": "Different code paths and purposes: loss scaling in training_step vs. column/signature filtering for model inputs. They would not plausibly be merged as one PR." } ] }, @@ -17691,52 +17652,52 @@ "pull_request:43656|pull_request:43842", "pull_request:43836|pull_request:43842" ], - "prepared_review_unit_hash": "1b3dc0ba6da7dd1bbec9b2b52eaa64f0839ce5ca8b062e4c7194cbbbf6b9f2d4", + "prepared_review_unit_hash": "8ba016a4f08f2fc6bb99862c3c11efb6636bcedb519e1440e12db066a340f207", "component_budget": { "node_count": 3, "item_count": 3, "soft_pair_count": 2, - "serialized_chars": 5991, - "estimated_input_tokens": 1498, - "estimated_eval_tokens": 3252 + "serialized_chars": 5993, + "estimated_input_tokens": 1499, + "estimated_eval_tokens": 3254 }, "budget": { "node_count": 3, "item_count": 3, "soft_pair_count": 2, - "serialized_chars": 5991, - "estimated_input_tokens": 1498, - "estimated_eval_tokens": 3252 + "serialized_chars": 5993, + "estimated_input_tokens": 1499, + "estimated_eval_tokens": 3254 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These PRs cluster around a CLI `serve.py` annotation/runtime NameError issue involving `TypeAdapter` and related typing cleanup. The strongest duplicate pair is the two explicit `TypeAdapter`-focused fixes; the more generic annotation cleanup looks related but not clearly the same concrete change.", - "confidence": 0.67, + "summary": "All three PRs target the same import-time TypeAdapter annotation problem in `src/transformers/cli/serve.py`; PR 43842 is the cleanest, most focused fix.", + "confidence": 0.86, "canonical_issue_reason": null, - "canonical_pr_reason": "PR #43842 is the most precise and narrowly scoped fix for the reported `TypeAdapter` NameError when `pydantic` is absent, with a minimal change set in the affected code path.", + "canonical_pr_reason": "PR 43842 is the most focused and explicit fix for the underlying NameError when pydantic is not installed, with a minimal annotation/import change in `serve.py`.", "best_issue_reason": null, - "best_pr_reason": "PR #43842 is the best representative because it directly targets the specific runtime failure, uses the smallest focused patch, and is least likely to include unrelated typing cleanup.", + "best_pr_reason": "PR 43842 best represents the cluster because it directly and narrowly fixes the concrete runtime NameError path without extra unrelated changes.", "soft_edge_verdicts": [ { "left": "pull_request:43656", "right": "pull_request:43842", - "accept": false, - "reason": "Both touch `src/transformers/cli/serve.py` and relate to annotation-name errors, but #43656 is broader future-annotations/type-hint cleanup while #43842 specifically fixes `TypeAdapter` availability; not clearly the same concrete bug fix." + "accept": true, + "reason": "Both PRs address the same import-time NameError from evaluating `TypeAdapter` in `src/transformers/cli/serve.py`; they are compatible fixes for the same code path." }, { "left": "pull_request:43836", "right": "pull_request:43842", "accept": true, - "reason": "Both are clearly aimed at the same `TypeAdapter` NameError in `serve.py`, using stringized annotations/TYPE_CHECKING to avoid runtime evaluation; these look like overlapping fixes to the same underlying issue." + "reason": "Both fixes target the same `TypeAdapter` annotation NameError in the CLI serve path; 43836 is broader, but its core change is the same bug fix and could plausibly be merged together." } ] }, "evaluator_result": { "accept": true, - "feedback": "The clustering is grounded in the packet. The distinction between 43656 and 43842 is conservative, and the 43836/43842 pairing is reasonably supported by the shared `serve.py` TypeAdapter fix and shared issue target. The summary does not overclaim beyond the evidence." + "feedback": "The summary is grounded in the packet: all three PRs touch `src/transformers/cli/serve.py` and are clearly aimed at preventing a `TypeAdapter`/annotation-related import-time NameError. The choice of PR 43842 as the most focused fix is supported by the smaller diff and explicit pydantic gating. The soft-edge accepts are also conservative enough: both pairs appear to be alternative or complementary fixes for the same bug path in the same file, so they look mergeable into a single bugfix PR." }, "evaluator_used": true, "retried": false, @@ -17862,8 +17823,8 @@ "pull_request:44066|pull_request:44086", "pull_request:44066|pull_request:44071", "pull_request:44018|pull_request:44068", - "pull_request:44066|pull_request:44068", "pull_request:44018|pull_request:44066", + "pull_request:44066|pull_request:44068", "pull_request:44019|pull_request:44722", "pull_request:44086|pull_request:44722", "pull_request:44019|pull_request:44085", @@ -17882,8 +17843,8 @@ "pull_request:44098|pull_request:44101", "pull_request:44059|pull_request:44722", "pull_request:44018|pull_request:44071", - "pull_request:44017|pull_request:44161", "pull_request:44017|pull_request:44722", + "pull_request:44017|pull_request:44161", "pull_request:44017|pull_request:44073", "pull_request:44068|pull_request:44071", "pull_request:44071|pull_request:44086", @@ -17892,8 +17853,8 @@ "pull_request:44073|pull_request:44101", "pull_request:44059|pull_request:44071", "pull_request:44056|pull_request:44101", - "pull_request:44073|pull_request:44161", "pull_request:44059|pull_request:44068", + "pull_request:44073|pull_request:44161", "pull_request:44018|pull_request:44059", "pull_request:44101|pull_request:44161", "pull_request:43997|pull_request:44017", @@ -17925,367 +17886,385 @@ "pull_request:44044|pull_request:44129", "pull_request:44044|pull_request:44154" ], - "prepared_review_unit_hash": "c7f05932f627e69748760f023cc4a63ce77ec28a38aee79e101cf9363479dc37", + "prepared_review_unit_hash": "933fa393f4ee6facf160dcc9f4a64f213051fc64cfba4ee53093a432823501f9", "component_budget": { "node_count": 28, "item_count": 28, "soft_pair_count": 78, - "serialized_chars": 84891, - "estimated_input_tokens": 21223, - "estimated_eval_tokens": 42702 + "serialized_chars": 84895, + "estimated_input_tokens": 21224, + "estimated_eval_tokens": 42704 }, "budget": { "node_count": 27, "item_count": 27, "soft_pair_count": 72, - "serialized_chars": 75088, - "estimated_input_tokens": 18772, - "estimated_eval_tokens": 37800 + "serialized_chars": 75093, + "estimated_input_tokens": 18774, + "estimated_eval_tokens": 37804 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly a batch of model-specific refactors for standardized output tracing around issue #43979. The only clear duplicate-like overlaps are within the GPT-Neo pair and the GPT-J/CodeGen sequence; the rest are separate model refactors that share a theme but not the same concrete change.", - "confidence": 0.82, + "summary": "Mostly independent per-model output-tracing refactors tied to #43979. There are at least two concrete mergeable overlap pockets: GPT-Neo (44018/44068) and GPT-J/CodeGen (44066/44085/44722), but the rest are separate model-specific changes.", + "confidence": 0.9, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 44722 is the best canonical representative for the concrete GPT-J/CodeGen output-tracing refactor: it has the clearest implementation scope, explicit issue linkage, and subsumes the closely related GPT-J-only variants.", + "canonical_pr_reason": "PR 44066 is a reasonable cluster anchor because it is an explicit #43979-linked GPT-J/CodeGen output-tracing refactor and overlaps with the later GPT-J/CodeGen follow-up, but it should be treated as one representative pocket rather than the sole center of the cluster.", "best_issue_reason": null, - "best_pr_reason": "PR 44722 is the strongest single PR in this cluster because it most clearly captures the underlying GPT-J/CodeGen output-tracing change and is the most complete version of that specific refactor.", + "best_pr_reason": "PR 44066 is the strongest single representative for the cluster because it has an explicit tracking-issue link and covers a broad concrete GPT-J/CodeGen refactor, while the cluster also contains a separate GPT-Neo overlap that should not be collapsed into the same anchor.", "soft_edge_verdicts": [ { "left": "pull_request:43996", "right": "pull_request:44085", "accept": false, - "reason": "Different model families and code paths; same issue target is not enough to make these duplicates." + "reason": "Same output-tracing initiative, but different model implementations and code paths." }, { "left": "pull_request:43996", "right": "pull_request:44044", "accept": false, - "reason": "CVT/FNet output tracing vs DeBERTa output tracing are separate model refactors." + "reason": "Both target #43979, but they are different model-specific refactors, not the same change." }, { "left": "pull_request:44066", "right": "pull_request:44085", "accept": true, - "reason": "Both touch the GPT-J output-tracing refactor on the same code path; 44066 just includes an added CodeGen companion change." + "reason": "Both touch the same GPT-J output-tracing code path, and these changes could plausibly be merged into one PR." }, { "left": "pull_request:44007", "right": "pull_request:44072", "accept": false, - "reason": "Different vision model refactors (ResNet/RT-DETR ResNet vs EfficientNet), not the same concrete change." + "reason": "Different model families and concrete code paths; only the broad output-tracing theme overlaps." }, { "left": "pull_request:44072", "right": "pull_request:44722", "accept": false, - "reason": "EfficientNet output tracing and GPT-J/CodeGen output tracing are unrelated concrete changes." + "reason": "Different model implementations; shared tracking-issue context is not enough to make this a duplicate." }, { "left": "pull_request:44066", "right": "pull_request:44072", "accept": false, - "reason": "GPT-J/CodeGen refactor vs EfficientNet refactor; only the shared theme is standardized tracing." + "reason": "Different models and different output-tracing code paths." }, { "left": "pull_request:44013", "right": "pull_request:44044", "accept": false, - "reason": "MobileNetV2 and DeBERTa are unrelated model implementations with different refactor targets." + "reason": "Different model families and unrelated concrete refactors." }, { "left": "pull_request:44066", "right": "pull_request:44086", "accept": false, - "reason": "GPT-J/CodeGen output tracing and MGP-STR output tracing are different code paths." + "reason": "Both are output-tracing work, but they target different model code paths." }, { "left": "pull_request:44066", "right": "pull_request:44071", "accept": false, - "reason": "GPT-J/CodeGen refactor vs MPT refactor; same theme, different model behavior." + "reason": "Different model implementations; same general refactor theme only." }, { "left": "pull_request:44018", "right": "pull_request:44068", "accept": true, - "reason": "These are both GPT-Neo output-tracing refactors of the same model code path and look like the same underlying change." + "reason": "These are the same GPT-Neo output-tracing refactor in the same model/file family and look mergeable as one PR." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44066", + "accept": false, + "reason": "GPT-Neo vs GPT-J/CodeGen are different concrete code paths." }, { "left": "pull_request:44066", "right": "pull_request:44068", "accept": false, - "reason": "GPT-J/CodeGen and GPT-Neo are separate model refactors, so this is only thematic similarity." + "reason": "GPT-J/CodeGen vs GPT-Neo are separate model-specific changes." }, { "left": "pull_request:44019", "right": "pull_request:44722", "accept": false, - "reason": "ResNet output tracing and GPT-J/CodeGen output tracing are unrelated changes." + "reason": "Different model families and different refactors." }, { "left": "pull_request:44086", "right": "pull_request:44722", "accept": false, - "reason": "MGP-STR and GPT-J/CodeGen are different model-specific refactors." + "reason": "Different model implementations; the shared theme is too broad to merge them." }, { "left": "pull_request:44019", "right": "pull_request:44085", "accept": false, - "reason": "ResNet and GPT-J are unrelated code paths despite the shared tracing theme." + "reason": "Different model families and output paths." }, { "left": "pull_request:44018", "right": "pull_request:44019", "accept": false, - "reason": "GPT-Neo and ResNet are different model implementations with distinct output handling." + "reason": "GPT-Neo and ResNet are unrelated concrete changes." }, { "left": "pull_request:44019", "right": "pull_request:44071", "accept": false, - "reason": "ResNet vs MPT is not the same underlying bug/change." + "reason": "Different models and different output-tracing code paths." }, { "left": "pull_request:44068", "right": "pull_request:44116", "accept": false, - "reason": "GPT-Neo output tracing and Flaubert output tracing are separate model changes." + "reason": "Different model families and unrelated concrete changes." }, { "left": "pull_request:44068", "right": "pull_request:44722", "accept": false, - "reason": "GPT-Neo and GPT-J/CodeGen are distinct model refactors, not one duplicate." + "reason": "Different model code paths; same initiative only." }, { "left": "pull_request:44071", - "right": "pull_request:44722", + "right": "pull_request:44116", "accept": false, - "reason": "MPT and GPT-J/CodeGen are different concrete output-tracing changes." + "reason": "MPT and ImageGPT are different model implementations." }, { "left": "pull_request:44018", - "right": "pull_request:44116", + "right": "pull_request:44086", "accept": false, - "reason": "GPT-Neo and Flaubert are separate models and separate changes." + "reason": "Different models; shared tracing pattern is too broad to merge." }, { "left": "pull_request:44018", - "right": "pull_request:44086", + "right": "pull_request:44085", "accept": false, - "reason": "GPT-Neo refactor vs MGP-STR refactor; same pattern, different code." + "reason": "GPT-Neo and GPT-J are separate model code paths." }, { - "left": "pull_request:44018", + "left": "pull_request:44071", "right": "pull_request:44085", "accept": false, - "reason": "GPT-Neo and GPT-J are different model refactors." + "reason": "Different models; only the generic output-tracing theme overlaps." }, { - "left": "pull_request:44098", - "right": "pull_request:44101", + "left": "pull_request:44068", + "right": "pull_request:44086", "accept": false, - "reason": "ViLT and Flaubert output-handling changes are unrelated concrete changes." + "reason": "Different models and different concrete refactors." }, { "left": "pull_request:44059", "right": "pull_request:44086", "accept": false, - "reason": "GPT2 and MGP-STR output tracing are separate code paths." + "reason": "GPT-2 vs MGP-STR are unrelated concrete changes." }, { "left": "pull_request:44085", "right": "pull_request:44086", "accept": false, - "reason": "GPT-J and MGP-STR are unrelated model implementations." + "reason": "Different models and code paths." + }, + { + "left": "pull_request:44098", + "right": "pull_request:44101", + "accept": false, + "reason": "Different model families and different fixes." }, { "left": "pull_request:44059", "right": "pull_request:44722", "accept": false, - "reason": "GPT2 refactor and GPT-J/CodeGen refactor are not the same bug/change." + "reason": "Different models; same theme is not enough for duplicate triage." }, { "left": "pull_request:44018", "right": "pull_request:44071", "accept": false, - "reason": "GPT-Neo vs MPT; only the broad tracing theme is shared." + "reason": "GPT-Neo and MPT are different model implementations." + }, + { + "left": "pull_request:44017", + "right": "pull_request:44722", + "accept": false, + "reason": "Different model families and concrete output paths." }, { "left": "pull_request:44017", "right": "pull_request:44161", "accept": false, - "reason": "SegFormer and LongT5 are different model-specific output-tracing refactors." + "reason": "Different model-specific refactors." }, { "left": "pull_request:44017", - "right": "pull_request:44722", + "right": "pull_request:44073", "accept": false, - "reason": "SegFormer and GPT-J/CodeGen are unrelated concrete changes." + "reason": "Different model families and code paths." }, { "left": "pull_request:44068", "right": "pull_request:44071", "accept": false, - "reason": "GPT-Neo and MPT are separate models with different implementation details." + "reason": "Different model implementations; same tracing pattern only." }, { "left": "pull_request:44071", - "right": "pull_request:44076", + "right": "pull_request:44086", "accept": false, - "reason": "MPT and ImageGPT are different model refactors." + "reason": "Different models and output implementations." }, { "left": "pull_request:43997", "right": "pull_request:44073", "accept": false, - "reason": "RegNet and VisualBert are unrelated model-specific changes." + "reason": "RegNet and VisualBert are unrelated concrete changes." }, { "left": "pull_request:44017", - "right": "pull_request:44071", + "right": "pull_request:44018", "accept": false, - "reason": "SegFormer vs MPT is only a shared architecture pattern, not the same change." + "reason": "Different model families and output paths." }, { - "left": "pull_request:44073", - "right": "pull_request:44101", + "left": "pull_request:44017", + "right": "pull_request:44056", "accept": false, - "reason": "VisualBert and Flaubert are different models and different output tracing updates." + "reason": "Different model families; same tracking initiative only." }, { - "left": "pull_request:44059", - "right": "pull_request:44071", + "left": "pull_request:44071", + "right": "pull_request:44129", "accept": false, - "reason": "GPT2 and MPT changes are unrelated concrete refactors." + "reason": "Different models and different output-tracing implementations." + }, + { + "left": "pull_request:44074", + "right": "pull_request:44101", + "accept": false, + "reason": "TextNet and Flaubert are unrelated concrete changes." }, { "left": "pull_request:44056", "right": "pull_request:44161", "accept": false, - "reason": "MPNet and LongT5 output tracing are separate code paths." + "reason": "Different models and refactor targets." }, { "left": "pull_request:43997", "right": "pull_request:44018", "accept": false, - "reason": "RegNet and GPT-Neo are unrelated implementations." + "reason": "RegNet and GPT-Neo are unrelated concrete code paths." }, { - "left": "pull_request:44017", - "right": "pull_request:44056", + "left": "pull_request:43997", + "right": "pull_request:44161", "accept": false, - "reason": "SegFormer and MPNet are different model refactors." + "reason": "Different model families; only the broad output-tracing theme overlaps." }, { - "left": "pull_request:44010", - "right": "pull_request:44071", + "left": "pull_request:44071", + "right": "pull_request:44076", "accept": false, - "reason": "SqueezeBert and MPT are different concrete changes." + "reason": "MPT and ImageGPT are different model-specific changes." }, { - "left": "pull_request:44073", - "right": "pull_request:44161", + "left": "pull_request:44010", + "right": "pull_request:44071", "accept": false, - "reason": "VisualBert and LongT5 are separate model-specific output-tracing refactors." + "reason": "SqueezeBert and MPT are different model implementations." }, { - "left": "pull_request:43997", - "right": "pull_request:44056", + "left": "pull_request:44018", + "right": "pull_request:44129", "accept": false, - "reason": "RegNet and MPNet are unrelated code paths." + "reason": "Different models and output code paths." }, { - "left": "pull_request:43997", - "right": "pull_request:44161", + "left": "pull_request:44010", + "right": "pull_request:44073", "accept": false, - "reason": "RegNet and LongT5 are not the same underlying bug/change." + "reason": "Different models; not the same underlying change." }, { "left": "pull_request:44074", "right": "pull_request:44161", "accept": false, - "reason": "TextNet and LongT5 are different model refactors." + "reason": "TextNet vs LongT5 are unrelated concrete changes." }, { - "left": "pull_request:44018", - "right": "pull_request:44059", - "accept": false, - "reason": "GPT-Neo and GPT2 are separate model-specific changes." - }, - { - "left": "pull_request:44056", - "right": "pull_request:44071", - "accept": false, - "reason": "MPNet and MPT are different implementations and not duplicates." - }, - { - "left": "pull_request:44010", - "right": "pull_request:44018", + "left": "pull_request:43996", + "right": "pull_request:44161", "accept": false, - "reason": "SqueezeBert and GPT-Neo are unrelated concrete changes." + "reason": "Different model families and output paths." }, { - "left": "pull_request:43997", + "left": "pull_request:43996", "right": "pull_request:44017", "accept": false, - "reason": "RegNet and SegFormer are different model refactors." + "reason": "Different model families and concrete code paths." }, { "left": "pull_request:44010", "right": "pull_request:44044", "accept": false, - "reason": "SqueezeBert and DeBERTa are unrelated output-tracing changes." + "reason": "Different models; the shared tracing theme is too broad." }, { "left": "pull_request:44044", "right": "pull_request:44074", "accept": false, - "reason": "DeBERTa and TextNet are separate model-specific changes." + "reason": "Different models and different fixes." }, { - "left": "pull_request:44010", - "right": "pull_request:44073", - "accept": false, - "reason": "SqueezeBert and VisualBert are different code paths." + "left": "pull_request:44066", + "right": "pull_request:44722", + "accept": true, + "reason": "Same GPT-J/CodeGen output-tracing family with overlapping GPT-J code; they could plausibly be merged into one PR." }, { - "left": "pull_request:44044", + "left": "pull_request:44085", + "right": "pull_request:44722", + "accept": true, + "reason": "Both are GPT-J output-tracing changes on the same file/code path, so they look mergeable as one PR." + }, + { + "left": "pull_request:44085", "right": "pull_request:44129", "accept": false, - "reason": "DeBERTa and SpeechT5 are separate model refactors." + "reason": "GPT-J and SpeechT5 are unrelated model-specific changes." }, { - "left": "pull_request:44044", - "right": "pull_request:44154", + "left": "pull_request:43996", + "right": "pull_request:44007", "accept": false, - "reason": "DeBERTa and VITS are different concrete output-handling changes." + "reason": "Different model families and concrete output paths." }, { - "left": "pull_request:44056", - "right": "pull_request:44068", + "left": "pull_request:43996", + "right": "pull_request:44129", "accept": false, - "reason": "MPNet and GPT-Neo are unrelated model refactors." + "reason": "Different model families; same initiative only." }, { - "left": "pull_request:44018", - "right": "pull_request:44071", + "left": "pull_request:44044", + "right": "pull_request:44129", "accept": false, - "reason": "GPT-Neo and MPT are different concrete changes." + "reason": "DeBERTa and SpeechT5 are unrelated concrete changes." } ] }, "evaluator_result": { - "accept": true, - "feedback": "The summary is grounded in the packet: most PRs are distinct model-specific output-tracing refactors, and the GPT-Neo and GPT-J/CodeGen overlaps are the only clear mergeable duplicates. The chosen canonical PR rationale is also supported by the diffs and issue linkage. Soft-edge verdicts are mostly conservative and do not overstate duplication." + "accept": false, + "feedback": "The packet supports the general conclusion that this is mostly a set of separate model-by-model output-tracing refactors tied to #43979, but the reasoning overstates the evidence in a few places. In particular, PR 44066 is not the center of the only clearly mergeable overlap: the GPT-Neo pair (44018/44068) is also a strong mergeable overlap, and 44066 also overlaps with 44722. Please keep the verdicts and anchor language more conservative and avoid implying uniqueness where the packet shows multiple mergeable pairs." }, "evaluator_used": true, - "retried": false, + "retried": true, "accepted_nontrivial_soft_edge": true, "error_kind": null, "error_message": null @@ -18320,20 +18299,20 @@ "pull_request:43997|pull_request:44010", "pull_request:44010|pull_request:44129" ], - "prepared_review_unit_hash": "0869a19b82fbb11817c2090a714046ed5347e7b46aae7422b0a476f28958921a", + "prepared_review_unit_hash": "edcafb15aa209634d5a49ca54432fc8d284926878e6f33cb4b2ae396bda186ab", "component_budget": { "node_count": 28, "item_count": 28, "soft_pair_count": 78, - "serialized_chars": 84891, - "estimated_input_tokens": 21223, - "estimated_eval_tokens": 42702 + "serialized_chars": 84895, + "estimated_input_tokens": 21224, + "estimated_eval_tokens": 42704 }, "budget": { "node_count": 8, "item_count": 8, "soft_pair_count": 6, - "serialized_chars": 17124, + "serialized_chars": 17123, "estimated_input_tokens": 4281, "estimated_eval_tokens": 8818 }, @@ -18342,48 +18321,48 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "These PRs are all instances of the same broad refactor pattern\u2014migrating individual model forwards to standardized output-capture utilities\u2014but they target different model implementations and code paths. As a duplicate cluster, they should stay separate; none of the soft edges are true duplicates under a strict PR triage standard.", - "confidence": 0.93, + "summary": "These PRs are related only by a broad theme: migrating different model implementations to the new standardized output-capture machinery. Each one targets a different model/code path, so they are not duplicates of the same underlying change.", + "confidence": 0.95, "canonical_issue_reason": null, - "canonical_pr_reason": "PR #44129 is the strongest cluster representative because it has the clearest problem statement, an explicit issue link, and a substantial output-tracing refactor in a model file consistent with the rest of the cluster.", + "canonical_pr_reason": "No single PR is a safe canonical duplicate target here: every candidate is model-specific and changes a different forward path, even though the refactor pattern is similar.", "best_issue_reason": null, - "best_pr_reason": "PR #44129 is the best single PR to represent the cluster since it most explicitly frames the standardized output-capture refactor and is easier to map back to the underlying initiative than the more model-specific variants.", + "best_pr_reason": "PR 44073 is the clearest representative of the shared refactor pattern, with explicit `capture_outputs` and `can_return_tuple` integration and a straightforward output-tracing migration, but it remains specific to VisualBert rather than a cluster-wide duplicate target.", "soft_edge_verdicts": [ { "left": "pull_request:44024", "right": "pull_request:44076", "accept": false, - "reason": "Same refactor theme, but different models and different code paths. These would not plausibly be merged into one PR without becoming an unrelated multi-model sweep." + "reason": "Both are standardized-output refactors, but they touch different models (FocalNet vs ImageGPT) and different forward paths; similar theme, not the same fix." }, { "left": "pull_request:44073", "right": "pull_request:44074", "accept": false, - "reason": "Both are output-tracing refactors, but they touch different architectures and files. Similar mechanics are not enough to make them duplicates." + "reason": "Different model implementations (VisualBert vs TextNet) with separate output-capture plumbing; they are not plausibly mergeable into one PR." }, { "left": "pull_request:44056", "right": "pull_request:44073", "accept": false, - "reason": "MPNet and VisualBert are separate model implementations; the shared decorator/capture approach is a framework pattern, not the same underlying bug or change." + "reason": "Both adopt output-tracing decorators, but they affect different architectures and code paths (MPNet vs VisualBert), so this is not the same underlying change." }, { "left": "pull_request:44056", "right": "pull_request:44074", "accept": false, - "reason": "Distinct model-specific refactors with no shared concrete defect or identical code path." + "reason": "MPNet and TextNet are unrelated model implementations; the shared refactor pattern is too broad to treat as a duplicate." }, { "left": "pull_request:43997", "right": "pull_request:44010", "accept": false, - "reason": "RegNet and SqueezeBert are unrelated model files. This is the same migration style, not the same change." + "reason": "RegNet and SqueezeBert are distinct model stacks with different forward logic; the overlap is only in migration style, not in the same bug or change." }, { "left": "pull_request:44010", "right": "pull_request:44129", "accept": false, - "reason": "These target different model internals and do not fix the same concrete issue; they only share the standardized output-capture motif." + "reason": "SqueezeBert and SpeechT5 modify different subsystems and output contracts; they share the decorator migration theme but not a single concrete code-path fix." } ] }, @@ -18463,274 +18442,274 @@ "pull_request:44028|pull_request:44030", "pull_request:44029|pull_request:44030" ], - "prepared_review_unit_hash": "55567ac97fae8f84953354214a32da6800b280f4fe44ad1b21474baebfb34a32", + "prepared_review_unit_hash": "fac27c8c4c121b855ebaf4ae33eb25c30a640387c3fffe83d14868c548661962", "component_budget": { "node_count": 13, "item_count": 13, "soft_pair_count": 40, - "serialized_chars": 40172, - "estimated_input_tokens": 10043, - "estimated_eval_tokens": 20342 + "serialized_chars": 40167, + "estimated_input_tokens": 10042, + "estimated_eval_tokens": 20340 }, "budget": { "node_count": 13, "item_count": 13, "soft_pair_count": 40, - "serialized_chars": 40172, - "estimated_input_tokens": 10043, - "estimated_eval_tokens": 20342 + "serialized_chars": 40167, + "estimated_input_tokens": 10042, + "estimated_eval_tokens": 20340 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These PRs share a common refactor theme, but they are separate model-specific changes rather than duplicates of one underlying bug or patch. I would not merge any soft edge pairs.", - "confidence": 0.95, + "summary": "This cluster is a set of similar but independent model-specific output-tracing refactors; they share a theme, not a single underlying bug or change.", + "confidence": 0.88, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 44003 is the most representative cluster anchor: it is a broader output-tracing refactor, spans two closely related model files, and introduces shared tracing helpers rather than a one-off local tweak.", + "canonical_pr_reason": "44025 is the best cluster anchor: it\u2019s a straightforward output-tracing refactor with a clear model-specific scope and touches two closely related depth_anything files.", "best_issue_reason": null, - "best_pr_reason": "PR 44003 is the best overall representative because it is the broadest and most infrastructure-oriented of the set, making it a better canonical example than the single-model refactors.", + "best_pr_reason": "44025 is the most representative PR in the set, with a clean refactor scope and the strongest centrality among the similar output-tracing changes.", "soft_edge_verdicts": [ { "left": "pull_request:43998", "right": "pull_request:44002", "accept": false, - "reason": "Different model families and files; same refactor theme only, not the same concrete bug/change." + "reason": "Different models and code paths; same refactor pattern, but not one mergeable change." }, { "left": "pull_request:43998", "right": "pull_request:44000", "accept": false, - "reason": "Different models/code paths; only the output-tracing refactor pattern overlaps." + "reason": "Different model families; this is broad output-tracing cleanup, not the same underlying fix." }, { "left": "pull_request:43998", "right": "pull_request:44001", "accept": false, - "reason": "Unrelated model implementation; no shared underlying fix." + "reason": "Timm backbone vs UnivNet are unrelated implementations; shared theme is too broad." }, { "left": "pull_request:43998", "right": "pull_request:44003", "accept": false, - "reason": "Different architectures and code paths; not one mergeable fix." + "reason": "Different subsystems and model stacks; no single concrete bug/change here." }, { "left": "pull_request:43998", "right": "pull_request:44029", "accept": false, - "reason": "Different model families; same tracing refactor motif but no duplicate bug." + "reason": "Different model code paths; similar tracing refactor only." }, { "left": "pull_request:43999", "right": "pull_request:44001", "accept": false, - "reason": "Separate models with different forward paths; not the same change." + "reason": "MobileNetV1 and UnivNet are unrelated models; not the same PR-worthy change." }, { "left": "pull_request:43999", "right": "pull_request:44002", "accept": false, - "reason": "Different subsystems and model code paths; only a generic refactor similarity." + "reason": "Different model implementations; too broad to treat as duplicates." }, { "left": "pull_request:43999", "right": "pull_request:44025", "accept": false, - "reason": "Different model families and filenames; not the same concrete fix." + "reason": "Different model families; same output-tracing theme but separate changes." }, { "left": "pull_request:43999", "right": "pull_request:44026", "accept": false, - "reason": "Encoder-decoder wrapper vs MobileNet model; no shared underlying bug." + "reason": "Different code paths and model families; not one concrete fix." }, { "left": "pull_request:43999", "right": "pull_request:44027", "accept": false, - "reason": "Different model code paths; same refactor pattern only." + "reason": "Model-specific refactors only; these are not mergeable as one PR." }, { "left": "pull_request:44000", "right": "pull_request:44001", "accept": false, - "reason": "Different models and output paths; not a duplicate change." + "reason": "Vision-text dual encoder vs UnivNet are unrelated; only the refactor pattern matches." }, { "left": "pull_request:44000", "right": "pull_request:44002", "accept": false, - "reason": "Different architectures and files; no same concrete bug/change." + "reason": "Different models and behavior; no shared underlying bug." }, { "left": "pull_request:44000", "right": "pull_request:44003", "accept": false, - "reason": "Distinct model families; too broad a similarity to merge." + "reason": "Different code paths and architectures; same theme, not same change." }, { "left": "pull_request:44000", "right": "pull_request:44025", "accept": false, - "reason": "Different model implementations and code paths; only same refactor style." + "reason": "Separate model implementations; output-tracing refactor is too generic to merge." }, { "left": "pull_request:44001", "right": "pull_request:44002", "accept": false, - "reason": "Separate model code paths; not the same underlying issue." + "reason": "UnivNet and UperNet are unrelated model paths; not the same fix." }, { "left": "pull_request:44001", "right": "pull_request:44003", "accept": false, - "reason": "Mamba/FalconMamba vs UnivNet; unrelated concrete changes." + "reason": "Different architectures and files; similar cleanup only." }, { "left": "pull_request:44001", "right": "pull_request:44025", "accept": false, - "reason": "Different models and outputs; no shared bug to coalesce." + "reason": "Different models; same refactor style is not enough to duplicate." }, { "left": "pull_request:44002", "right": "pull_request:44029", "accept": false, - "reason": "Different model families and forward implementations; not duplicates." + "reason": "UperNet and RWKV are unrelated; no common concrete bug/change." }, { "left": "pull_request:44002", "right": "pull_request:44003", "accept": false, - "reason": "Distinct code paths in different architectures; same meta-theme only." + "reason": "Different model stacks and code paths; not one change." }, { "left": "pull_request:44003", "right": "pull_request:44028", "accept": false, - "reason": "Different model families and files; not the same fix." + "reason": "Mamba/Falcon-Mamba vs SuperPoint are unrelated implementations; only superficially similar." }, { "left": "pull_request:44003", "right": "pull_request:44004", "accept": false, - "reason": "Mamba/RWKV vs CodeGen are unrelated code paths; do not merge as duplicates." + "reason": "Different model families and forward signatures; not a single mergeable fix." }, { "left": "pull_request:43998", "right": "pull_request:44004", "accept": false, - "reason": "Different model subsystems; only the tracing refactor label matches." + "reason": "Timm backbone and CodeGen are unrelated; same refactor theme is too broad." }, { "left": "pull_request:44001", "right": "pull_request:44004", "accept": false, - "reason": "Separate model implementations; not one concrete bug." + "reason": "UnivNet vs CodeGen are different code paths; not duplicates." }, { "left": "pull_request:44002", "right": "pull_request:44004", "accept": false, - "reason": "Different model families and forward logic; not duplicates." + "reason": "Different models and behavior; no shared concrete bug." }, { "left": "pull_request:44000", "right": "pull_request:44004", "accept": false, - "reason": "Vision/text dual encoder vs CodeGen are unrelated code paths." + "reason": "Different model implementations; not the same underlying change." }, { "left": "pull_request:44025", "right": "pull_request:44026", "accept": false, - "reason": "Different model families; similar refactor pattern only." + "reason": "Different model wrappers and files; same tracing refactor pattern only." }, { "left": "pull_request:44025", "right": "pull_request:44027", "accept": false, - "reason": "DepthAnything vs SpeechEncoderDecoder are unrelated concrete changes." + "reason": "DepthAnything vs SpeechEncoderDecoder are unrelated; not one bug fix." }, { "left": "pull_request:44025", "right": "pull_request:44028", "accept": false, - "reason": "Different models and output structures; not the same fix." + "reason": "Different models and outputs; similar refactor, not duplicate." }, { "left": "pull_request:44025", "right": "pull_request:44029", "accept": false, - "reason": "Different architectures and files; no shared bug." + "reason": "DepthAnything and RWKV are unrelated code paths." }, { "left": "pull_request:44025", "right": "pull_request:44030", "accept": false, - "reason": "DepthAnything vs DPR are distinct changes in different code paths." + "reason": "Different model families; same output-tracing cleanup only." }, { "left": "pull_request:44026", - "right": "pull_request:44029", + "right": "pull_request:44027", "accept": false, - "reason": "Different encoder-decoder wrappers vs RWKV; not one underlying issue." + "reason": "Encoder-decoder wrappers for different modalities; not the same concrete change." }, { "left": "pull_request:44026", - "right": "pull_request:44027", + "right": "pull_request:44028", "accept": false, - "reason": "Different model wrappers and files; same tracing pattern only." + "reason": "VisionEncoderDecoder and SuperPoint are unrelated; only the tracing refactor theme overlaps." }, { "left": "pull_request:44026", - "right": "pull_request:44028", + "right": "pull_request:44029", "accept": false, - "reason": "Different models and concrete forward paths; not duplicates." + "reason": "Different model stacks and behavior; not mergeable as one PR." }, { "left": "pull_request:44026", "right": "pull_request:44030", "accept": false, - "reason": "Vision/audio encoder-decoder vs DPR are unrelated code changes." + "reason": "Different model code paths; similar cleanup, not duplicate." }, { "left": "pull_request:44027", "right": "pull_request:44028", "accept": false, - "reason": "Different models and implementations; no shared bug/change." + "reason": "SpeechEncoderDecoder vs SuperPoint are unrelated implementations." }, { "left": "pull_request:44027", "right": "pull_request:44029", "accept": false, - "reason": "Speech encoder-decoder and RWKV are distinct code paths." + "reason": "Different model families and forward logic; not the same bug/change." }, { "left": "pull_request:44027", "right": "pull_request:44030", "accept": false, - "reason": "Different models and output mechanisms; not mergeable as duplicates." + "reason": "Separate model-specific refactors; not one concrete PR." }, { "left": "pull_request:44028", "right": "pull_request:44029", "accept": false, - "reason": "SuperPoint vs RWKV are unrelated changes despite similar refactor wording." + "reason": "SuperPoint and RWKV are unrelated; shared refactor pattern is insufficient." }, { "left": "pull_request:44028", "right": "pull_request:44030", "accept": false, - "reason": "Different model families and code paths; not the same fix." + "reason": "Different model code paths; not the same underlying change." }, { "left": "pull_request:44029", "right": "pull_request:44030", "accept": false, - "reason": "RWKV and DPR are different implementations; only the output-tracing refactor theme overlaps." + "reason": "RWKV and DPR are separate models; no single mergeable fix." } ] }, @@ -18763,12 +18742,12 @@ "pull_request:44662|pull_request:45075", "pull_request:45075|pull_request:45401" ], - "prepared_review_unit_hash": "f924fe59d79e0dce6c7f557453056dd024ef64ccba3a6dc6bde8c7912eef62d9", + "prepared_review_unit_hash": "379b431497882f505caa509e54d6408b81cbd5ce2441305d2c43838eb19b40b6", "component_budget": { "node_count": 3, "item_count": 3, "soft_pair_count": 3, - "serialized_chars": 10913, + "serialized_chars": 10914, "estimated_input_tokens": 2729, "estimated_eval_tokens": 5714 }, @@ -18776,7 +18755,7 @@ "node_count": 3, "item_count": 3, "soft_pair_count": 3, - "serialized_chars": 10913, + "serialized_chars": 10914, "estimated_input_tokens": 2729, "estimated_eval_tokens": 5714 }, @@ -18785,30 +18764,30 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are three separate model-addition PRs (PenguinVL, DeepSeek-OCR-2, and Voxtral TTS). They share some Transformers registry/docs plumbing, but they do not appear to fix the same underlying bug or implement the same concrete change, so none should be deduplicated together.", - "confidence": 0.94, - "canonical_issue_reason": null, - "canonical_pr_reason": "PR #45075 is the best cluster representative: it is a substantial, non-draft model integration with conversion code, modeling, processing, tests, and docs. The other PRs are distinct model additions and not duplicates.", - "best_issue_reason": null, - "best_pr_reason": "PR #45075 is the strongest standalone PR in this set because it looks the most complete and self-contained, with explicit conversion and broad implementation coverage.", + "summary": "These PRs all add separate new model integrations, but for different models and code paths, so they are not duplicates.", + "confidence": 0.98, + "canonical_issue_reason": "No issue items are present in this cluster.", + "canonical_pr_reason": "No canonical PR: PenguinVL, DeepSeek-OCR-2, and Voxtral TTS are distinct model additions, not the same change.", + "best_issue_reason": "No issue candidate to select as best issue.", + "best_pr_reason": "No single best PR for deduplication; each PR implements a different model integration.", "soft_edge_verdicts": [ { "left": "pull_request:44662", "right": "pull_request:45401", "accept": false, - "reason": "PenguinVL and Voxtral TTS are different models with different codepaths; overlap is limited to shared Transformers registry/docs files, which is not enough to treat them as the same change." + "reason": "PenguinVL and Voxtral TTS are different model additions with different architectures and functionality; they do not fix the same underlying problem and would not be merged as one PR." }, { "left": "pull_request:44662", "right": "pull_request:45075", "accept": false, - "reason": "PenguinVL and DeepSeek-OCR-2 are unrelated model implementations. Shared auto-registration and docs files are expected for separate additions, not evidence of duplication." + "reason": "PenguinVL and DeepSeek-OCR-2 are separate model implementations; shared registration/docs boilerplate is not enough to treat them as the same change." }, { "left": "pull_request:45075", "right": "pull_request:45401", "accept": false, - "reason": "DeepSeek-OCR-2 and Voxtral TTS target different architectures and functionality. They only share common model-registration plumbing, so they should not be merged as duplicates." + "reason": "DeepSeek-OCR-2 and Voxtral TTS add different models in different domains, so they are not the same code-path change or duplicate PR." } ] }, @@ -18838,12 +18817,12 @@ "soft_pairs": [ "pull_request:44979|pull_request:45363" ], - "prepared_review_unit_hash": "e30cbd0e07823ba3b76575d557f9fee68a9e9a003e77cfafdae5078174e677cf", + "prepared_review_unit_hash": "65e585d2f265df15d64ab74be5f85bc21c34ef0ed34ede50226c82a7ad2b027f", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4408, + "serialized_chars": 4405, "estimated_input_tokens": 1102, "estimated_eval_tokens": 2460 }, @@ -18851,7 +18830,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4408, + "serialized_chars": 4405, "estimated_input_tokens": 1102, "estimated_eval_tokens": 2460 }, @@ -18860,18 +18839,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "The two PRs are thematically related around fusion, but they target different layers: one introduces a general module-fusion API, while the other wires up n-to-1 kernel fusion through KernelConfig and model-loading integration. They are not duplicates of the same concrete change.", - "confidence": 0.91, + "summary": "These two PRs are related in the broad area of module/kernel fusion, but they are not the same change. PR #44979 adds a generic Module Fusion API and tests, while PR #45363 implements a specific n-to-1 kernel fusion workflow via KernelConfig across multiple integration/modeling files. The overlap is architectural, not a duplicate bug fix or mergeable single patch.", + "confidence": 0.86, "canonical_issue_reason": null, - "canonical_pr_reason": "PR #45363 is the more concrete end-user feature: it connects KernelConfig to model loading and kernel-backed fusion across multiple integration points. It is the better representative for the cluster\u2019s practical fusion functionality.", + "canonical_pr_reason": "#45363 is the more concrete end-user feature implementation: it wires fusion through KernelConfig, touches the actual integration/modeling path, and includes an example. #44979 is the underlying API scaffold, not the same deliverable.", "best_issue_reason": null, - "best_pr_reason": "PR #45363 is the strongest overall representative because it has the clearest feature surface, explicit integration changes, and a user-visible workflow. PR #44979 is more of a foundational API piece than the full feature.", + "best_pr_reason": "#45363 is the best representative PR for the cluster because it contains the practical fusion behavior and integration changes. #44979 is broader infrastructure, but it does not subsume the specific n-to-1 kernel fusion feature.", "soft_edge_verdicts": [ { "left": "pull_request:44979", "right": "pull_request:45363", "accept": false, - "reason": "Related area, but not the same underlying change. #44979 adds a standalone module fusion API in src/transformers/module_fusion.py; #45363 implements n-to-1 kernel fusion via KernelConfig and loader/integration plumbing. Different abstractions and code paths, so they should not be merged as duplicates." + "reason": "Do not treat as duplicates: #44979 introduces a generic module fusion API, whereas #45363 adds a specific KernelConfig-based n-to-1 kernel fusion implementation. They share a theme but not the same concrete code-path change." } ] }, @@ -18901,7 +18880,7 @@ "soft_pairs": [ "pull_request:45086|pull_request:45317" ], - "prepared_review_unit_hash": "4e3706aa6b9f0a8b7550d2da59a2e3e3f9abe18701f80fce9bbc8879c79d92e9", + "prepared_review_unit_hash": "24c7c535cdc03a35c40f44017f8591df820e088d770584d2ea6bbb47a11c8ad3", "component_budget": { "node_count": 2, "item_count": 2, @@ -18923,18 +18902,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Both pull requests fix the same AttributeError in `_patch_mistral_regex` by switching from `tokenizer.backend_tokenizer.pre_tokenizer` to `tokenizer.pre_tokenizer`; 45317 is the more complete version because it includes tests and clearer framing.", - "confidence": 0.98, + "summary": "Both PRs address the same Mistral regex patch AttributeError in the same code path; 45317 is the fuller version with tests, while 45086 is a smaller earlier variant.", + "confidence": 0.97, "canonical_issue_reason": null, - "canonical_pr_reason": "45317 is the better canonical PR because it targets the same bug as 45086 but adds test coverage and a clearer description of the fix.", + "canonical_pr_reason": "45317 is the better canonical PR because it targets the same fix as 45086 but adds test coverage and appears more complete.", "best_issue_reason": null, - "best_pr_reason": "45317 is the best representative PR for this cluster since it fully addresses the bug and includes validation, whereas 45086 is a smaller patch of the same change.", + "best_pr_reason": "45317 is the strongest representative of the cluster: same bug fix, same file, same issue target, plus tests verifying the behavior.", "soft_edge_verdicts": [ { "left": "pull_request:45086", "right": "pull_request:45317", "accept": true, - "reason": "Same underlying bug and same code path in `tokenization_utils_tokenizers.py`; the later PR is essentially a superset with added tests, so they could plausibly be merged into one fix." + "reason": "Same underlying AttributeError fix in _patch_mistral_regex; same file and issue target, and the later PR extends the same change with tests rather than a different fix." } ] }, @@ -18964,7 +18943,7 @@ "soft_pairs": [ "pull_request:45115|pull_request:45569" ], - "prepared_review_unit_hash": "7f8ced45f61375ec7d764c00d673fdfe23dc51164b968d0bde7e31ce10247fc6", + "prepared_review_unit_hash": "93db06790329c486fa5b959ba9141d15af12809135c950c560a918f329ed9929", "component_budget": { "node_count": 2, "item_count": 2, @@ -18986,18 +18965,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Related NemotronH work, but not duplicates: PR #45115 is a conversion/refactor for legacy NemotronH weight mapping and GraniteMoeHybrid-compatible naming, while PR #45569 is a broader feature split that adds NemotronH Dense/Sparse variants, auto-mapping updates, docs, and tests.", - "confidence": 0.93, + "summary": "Both PRs touch NemotronH internals, but they are different changes: one is a small conversion/mapping and inheritance refactor, the other is a broader architectural split into NemotronHDense/NemotronHSparse with docs and tests.", + "confidence": 0.89, "canonical_issue_reason": null, - "canonical_pr_reason": "PR #45569 is the broader umbrella change and the better representative of the NemotronH family update; it includes the new Dense/Sparse model structure and the surrounding integration work.", + "canonical_pr_reason": "PR 45569 is the better canonical PR because it represents the larger, more complete NemotronH restructuring, including the new dense/sparse model split, auto-mapping updates, docs, and tests.", "best_issue_reason": null, - "best_pr_reason": "PR #45569 is the most complete and user-visible change in the cluster, so it is the best single PR to anchor on if one must be chosen.", + "best_pr_reason": "PR 45569 is the best representative of the cluster since it is the more comprehensive user-facing/model-architecture change; PR 45115 is a narrower preparatory refactor focused on conversion mapping and inheritance wiring.", "soft_edge_verdicts": [ { "left": "pull_request:45115", "right": "pull_request:45569", "accept": false, - "reason": "They touch some of the same NemotronH files, but the underlying changes are different: #45115 is a conversion/mapping refactor for inheritance from GraniteMoeHybrid, while #45569 introduces Dense/Sparse model variants plus docs/tests. They would not plausibly merge as one duplicate PR." + "reason": "Reject: they share NemotronH-related files, but the concrete changes differ. 45115 is a focused conversion-mapping/refactor PR, while 45569 introduces the dense/sparse model split plus docs/tests. They are related work, not the same underlying change and would not plausibly be merged into one PR." } ] }, @@ -19027,7 +19006,7 @@ "soft_pairs": [ "pull_request:45332|pull_request:45333" ], - "prepared_review_unit_hash": "ec39b6b427c5404d80288dd59f606316fdede2caff9acbc8baac59d335402ae4", + "prepared_review_unit_hash": "37732b33b8b5144327a3eccb2d00d78a67834b5a8cf0064fca19fbe7241a90df", "component_budget": { "node_count": 2, "item_count": 2, @@ -19049,18 +19028,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Two related PRs in the heterogeneous-model workstream, but they are not the same change: #45333 adds per-layer heterogeneous config support, while #45332 builds on that with broader modeling/cache changes for heterogeneous models. They should stay separate.", - "confidence": 0.89, + "summary": "Two closely related but distinct PRs: one adds the heterogeneous per-layer config API, and the other builds model/cache support on top of it. They share infrastructure, but they are not the same concrete change.", + "confidence": 0.91, "canonical_issue_reason": null, - "canonical_pr_reason": "PR #45332 is the broader umbrella change: it adds heterogeneous model support across cache/modeling code and includes the downstream consumers of the new per-layer config abstractions.", + "canonical_pr_reason": "PR #45332 is the better canonical representative because it is the broader umbrella change: it adds heterogeneous model support and also includes the config plumbing it depends on. PR #45333 is a narrower prerequisite subset focused on config serialization/representation.", "best_issue_reason": null, - "best_pr_reason": "PR #45332 is the best representative of the cluster because it spans the full heterogeneous-model feature set, not just the config plumbing.", + "best_pr_reason": "PR #45332 is the strongest single representative of the cluster because it covers the larger end-to-end feature and incorporates the shared heterogeneity foundation used by #45333.", "soft_edge_verdicts": [ { "left": "pull_request:45332", "right": "pull_request:45333", "accept": false, - "reason": "Related work, but not the same concrete change. #45333 is config plumbing only; #45332 extends heterogeneous support into modeling/cache behavior and is broader than a mergeable duplicate." + "reason": "Do not treat these as duplicates: #45333 is config-layer support only, while #45332 adds broader model/cache integration and downstream behavior. They are related and sequenced, but not the same underlying code-path fix or mergeable into one PR without collapsing distinct scopes." } ] }, @@ -19090,7 +19069,7 @@ "soft_pairs": [ "pull_request:45415|pull_request:45425" ], - "prepared_review_unit_hash": "ef2d5f67b781a3ee50ac25c5911b63a0b1e43e5b3ae727afe5ed9244cde250ec", + "prepared_review_unit_hash": "764bb3966caa37bafe982d589372f28f8e0b1ee7b003c6edbec50b83d8754ad3", "component_budget": { "node_count": 2, "item_count": 2, @@ -19112,18 +19091,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Both PRs are typing-related, but they address different scopes: #45415 is a broad repository-wide type-checking/configuration pass, while #45425 adds specific typing protocols and helpers for model/PEFT-related code paths. They are related, but not the same change.", - "confidence": 0.91, + "summary": "Both items are typing-related pull requests, but they make different changes: one is a broad repo-wide type-checking cleanup, the other adds specific typing protocols and related model-utils annotations. They should not be treated as duplicates.", + "confidence": 0.93, "canonical_issue_reason": null, - "canonical_pr_reason": "PR #45425 is the more concrete and representative typing change: it adds specific protocol/type definitions used by model-related code, rather than a broad type-checking sweep.", + "canonical_pr_reason": "PR 45425 is the more focused, code-path-specific typing change, adding shared typing protocols and model-utils-related annotations; PR 45415 is a much broader type-checking sweep across many files and a config tweak, so it is not the best representative of the cluster.", "best_issue_reason": null, - "best_pr_reason": "PR #45425 is the best single representative artifact because it has a focused implementation goal and clear code-path impact, whereas #45415 is a wider type-checking cleanup across many files.", + "best_pr_reason": "PR 45425 is the better cluster representative because it introduces concrete typing infrastructure changes in core modules, whereas PR 45415 is mostly a wide annotation/checker cleanup touching many unrelated files.", "soft_edge_verdicts": [ { "left": "pull_request:45415", "right": "pull_request:45425", "accept": false, - "reason": "Do not treat as duplicates: #45415 is a broad type-checking pass/config change across many modules, while #45425 introduces specific typing helpers for model/PEFT integration code. Related theme, different underlying change." + "reason": "They overlap on typing/theme and share a few filenames, but the underlying changes are different: 45415 is a broad mypy/type-ignore cleanup across many files, while 45425 adds specific typing protocols and model-utils support. Not the same concrete fix/change." } ] }, @@ -19162,70 +19141,70 @@ "pull_request:45858|pull_request:45926", "pull_request:45890|pull_request:45926" ], - "prepared_review_unit_hash": "70f362b2fd0f75bfa7df5896e8e4a2b1e4ba3dc7143168794f4318e6bf2763c9", + "prepared_review_unit_hash": "f544d3c5395f37af9059cac4274a03ea3c9a0f0cf8fb4405a565e24bf3797892", "component_budget": { "node_count": 6, "item_count": 6, "soft_pair_count": 6, - "serialized_chars": 14704, - "estimated_input_tokens": 3676, - "estimated_eval_tokens": 7608 + "serialized_chars": 14717, + "estimated_input_tokens": 3680, + "estimated_eval_tokens": 7616 }, "budget": { "node_count": 6, "item_count": 6, "soft_pair_count": 6, - "serialized_chars": 14704, - "estimated_input_tokens": 3676, - "estimated_eval_tokens": 7608 + "serialized_chars": 14717, + "estimated_input_tokens": 3680, + "estimated_eval_tokens": 7616 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These PRs are mostly separate model-specific fixes or test expectation updates. They share a broad theme of parallel/XPU-related work, but not the same underlying bug/change, so none of the soft edges should be merged.", - "confidence": 0.95, + "summary": "This cluster mixes several unrelated PRs: two real model-parallel code fixes and three separate XPU test-expectation updates. None of the soft edges should be merged as duplicates.", + "confidence": 0.93, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 45817 is the strongest representative: it contains a concrete code fix for Gemma4 model-parallel behavior, with an actual implementation change rather than only test-data updates.", + "canonical_pr_reason": "No single PR is a true duplicate of the others, but #45487 is the best representative of the cluster\u2019s substantive bug-fix side because it changes model code to fix device placement in AltCLIP/ChineseCLIP rather than only updating tests.", "best_issue_reason": null, - "best_pr_reason": "PR 45817 is the best single PR to represent the cluster because it is a substantive bug fix in model-parallel logic and is more central than the small test-only PRs.", + "best_pr_reason": "#45487 is the strongest canonical PR candidate overall: it addresses an actual code-path bug in model parallel execution and spans the most directly related affected models, while the others are either different model families or test-only expectation updates.", "soft_edge_verdicts": [ { "left": "pull_request:45487", "right": "pull_request:45843", "accept": false, - "reason": "Both are model-parallel fixes, but for different models and different failure modes: AltCLIP/ChineseCLIP device placement vs Deimv2 _no_split_modules propagation. Not the same underlying bug." + "reason": "Both mention model parallel, but they fix different models and different code paths (AltCLIP/ChineseCLIP vs Deimv2). Not the same underlying bug." }, { "left": "pull_request:45487", "right": "pull_request:45926", "accept": false, - "reason": "AltCLIP/ChineseCLIP code fix versus HyperCLOVAX XPU expectation updates. Different models and one is test-only, so they should not be merged." + "reason": "#45487 is a source-code device placement fix; #45926 only adds XPU test expectations for HyperCLOVA X. Different models and different change types." }, { "left": "pull_request:45817", "right": "pull_request:45843", "accept": false, - "reason": "Gemma4 and Deimv2 have distinct model-parallel code paths and fixes. Same broad area, but not the same concrete bug." + "reason": "Both are model-parallel fixes, but for different architectures and distinct implementation issues. They would not plausibly be merged into one PR." }, { "left": "pull_request:45858", "right": "pull_request:45890", "accept": false, - "reason": "Both only add XPU expectations, but they target different model test suites and different outputs. No shared underlying bug/change." + "reason": "These are separate test-only XPU expectation updates for different models, not the same bug or change." }, { "left": "pull_request:45858", "right": "pull_request:45926", "accept": false, - "reason": "Both are test expectation updates, but for different models and test cases. Too broad to treat as duplicates." + "reason": "Different model test expectation additions for different architectures; similarity is only at the broad CI/test level." }, { "left": "pull_request:45890", "right": "pull_request:45926", "accept": false, - "reason": "These are separate XPU expectation updates for different model suites, not the same concrete fix or change." + "reason": "Separate model-specific XPU expectation updates; no shared concrete code-path problem." } ] }, @@ -19255,7 +19234,7 @@ "soft_pairs": [ "pull_request:45548|pull_request:45662" ], - "prepared_review_unit_hash": "d2a708fe0009cc9081a1359d6893ba957a0123cdcc1ce169989d2d9b2f4304fa", + "prepared_review_unit_hash": "c192077cd647c9d391cad276140a94397734170337ad7d3c84919bc5856bce08", "component_budget": { "node_count": 2, "item_count": 2, @@ -19277,18 +19256,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These two PRs both touch expert parallelism, but they address different failures: one is about preserving/loading EP-related config and enabling EP detection during DeepSpeed ZeRO-3 loading, while the other fixes FSDP2/rank-0 broadcast overwriting expert parameters during distributed loading. They are related subsystem-wise but not the same underlying bug or change.", - "confidence": 0.96, + "summary": "Both PRs are about expert-parallel (EP) loading in Transformers, but they address different failure modes and distributed backends: one is focused on accelerate + DeepSpeed ZeRO-3/config serialization, the other on FSDP2 rank-0 broadcast overwriting expert tensors. They are related but not the same underlying change.", + "confidence": 0.89, "canonical_issue_reason": null, - "canonical_pr_reason": null, + "canonical_pr_reason": "PR #45662 is the better canonical representative for the cluster because it targets the core loading path and fixes a concrete EP corruption bug for distributed tensor handling; #45548 is narrower and tied to accelerate/ZeRO-3 config serialization.", "best_issue_reason": null, - "best_pr_reason": "PR 45662 is the more concrete and self-contained bug fix: it targets a specific expert-parameter overwrite path during distributed loading and includes code changes directly addressing that failure mode.", + "best_pr_reason": "PR #45662 is the strongest standalone candidate: it fixes a specific, reproducible code-path problem in model loading and has the broader technical scope among the two. #45548 is more specialized and does not cover the same failure mode.", "soft_edge_verdicts": [ { "left": "pull_request:45548", "right": "pull_request:45662", "accept": false, - "reason": "Different concrete bugs and code paths: 45548 focuses on config serialization/EP detection for DeepSpeed ZeRO-3 loading, while 45662 fixes expert tensors being overwritten during FSDP2/loading broadcast. They should not be merged as duplicates." + "reason": "Related EP-loading area, but different bugs and backends: #45548 removes distributed_config from serialization and adds has_ep for accelerate/ZeRO-3, while #45662 prevents expert parameters from being overwritten after sharding in FSDP2. They would not naturally merge into one PR." } ] }, @@ -19318,12 +19297,12 @@ "soft_pairs": [ "pull_request:45639|pull_request:45645" ], - "prepared_review_unit_hash": "ad207c371f962bbdddaf4437f9f0e89b21e205efadafb7dbf26f966a21593228", + "prepared_review_unit_hash": "187a07454dfada8b44baf625e90d9b61819d8c1a2070073e97efa646878254e7", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4590, + "serialized_chars": 4589, "estimated_input_tokens": 1148, "estimated_eval_tokens": 2552 }, @@ -19331,7 +19310,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4590, + "serialized_chars": 4589, "estimated_input_tokens": 1148, "estimated_eval_tokens": 2552 }, @@ -19340,18 +19319,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Both PRs address the same xdist-related captured_info/debug-log collision problem in testing_utils; 45645 is the more complete variant because it also updates CI log collection to handle per-worker files.", - "confidence": 0.86, + "summary": "Both PRs address the same xdist-related captured_info logging problem; #45645 is the broader superset and likely the better canonical PR.", + "confidence": 0.81, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 45645 is the best canonical representative because it includes the core xdist-safe testing_utils fix plus the downstream CI workflow adjustment needed to consume multiple captured_info files.", + "canonical_pr_reason": "#45645 covers the full end-to-end fix for xdist-safe captured_info artifacts, including the testing utility change plus CI log consumption updates, so it is the more complete canonical PR.", "best_issue_reason": null, - "best_pr_reason": "PR 45645 is the strongest standalone candidate: it covers the same underlying bug as 45639 and adds the supporting CI handling, making it the more complete fix.", + "best_pr_reason": "#45645 is the best representative because it subsumes the core fix in #45639 and adds the related CI handling needed to preserve and read all captured_info files.", "soft_edge_verdicts": [ { "left": "pull_request:45639", "right": "pull_request:45645", "accept": true, - "reason": "Accept: both PRs fix the same captured_info/xdist collision in the test debugging path, and 45645 looks like a broader follow-up of the same concrete change rather than a different bug." + "reason": "Same underlying bug: xdist collisions in captured_info handling for patched testing debug logs. #45645 includes the same testing_utils/test changes and extends them with workflow-side log aggregation, so these look like one coherent fix that could plausibly be merged together." } ] }, @@ -19510,12 +19489,12 @@ "soft_pairs": [ "pull_request:45851|pull_request:45947" ], - "prepared_review_unit_hash": "2133b4f90d7396d57ff929fb3a603b0289caf70ccdb2be432cae7e415eb494d1", + "prepared_review_unit_hash": "0836afdc4169fa50f00f9ad52c4c9bc6611b7530d4d254c1bb2dcbdcb4bf1e37", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4105, + "serialized_chars": 4106, "estimated_input_tokens": 1027, "estimated_eval_tokens": 2310 }, @@ -19523,7 +19502,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4105, + "serialized_chars": 4106, "estimated_input_tokens": 1027, "estimated_eval_tokens": 2310 }, @@ -19532,18 +19511,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Two open docs PRs in the chat templating area, but they address different topics: template file storage/loading vs. reasoning-field prefill behavior. They should not be deduplicated.", - "confidence": 0.96, + "summary": "These are two separate docs-only pull requests in the chat templating area, but they address different topics: template file/storage docs vs. reasoning-field prefill behavior.", + "confidence": 0.93, "canonical_issue_reason": null, - "canonical_pr_reason": "PR #45851 is the better anchor because it is explicitly linked to issue #45205 and has a focused, self-contained docs change about chat template storage/loading.", + "canonical_pr_reason": "No canonical PR duplicate here; the patches are about different chat templating documentation changes rather than the same fix.", "best_issue_reason": null, - "best_pr_reason": "PR #45851 is the strongest representative: it has an explicit issue target, a clearer scoped documentation fix, and is more substantively connected to a concrete user-facing gap than #45947.", + "best_pr_reason": "PR 45851 is the better representative because it has an explicit issue target and a broader, more central documentation update.", "soft_edge_verdicts": [ { "left": "pull_request:45851", "right": "pull_request:45947", "accept": false, - "reason": "They are both chat-templating docs, but they describe different changes and different behaviors. #45851 explains on-disk chat template shapes; #45947 documents prefilling a reasoning field instead of content. Not the same underlying bug/change." + "reason": "Different changes and different user-facing concerns: one documents chat template storage/writing, the other documents prefill behavior for reasoning fields. They share a subsystem but not the same underlying change." } ] }, diff --git a/analysis/current/manifest.json b/analysis/current/manifest.json index 016f10706614afc40420230e9e3a09eb4953241f..924f2dc1427b54b9e1cc7a2bdc31c2a4665de998 100644 --- a/analysis/current/manifest.json +++ b/analysis/current/manifest.json @@ -1,8 +1,8 @@ { - "analysis_id": "hybrid-model-20260514t000044z", + "analysis_id": "hybrid-model-20260514t060044z", "archived_artifacts": { - "hybrid": "snapshots/20260514T000044Z/analysis-runs/hybrid-model-20260514t000044z/analysis-report-hybrid.json", - "hybrid_reviews": "snapshots/20260514T000044Z/analysis-runs/hybrid-model-20260514t000044z/analysis-report-hybrid.llm-reviews.json" + "hybrid": "snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.json", + "hybrid_reviews": "snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.llm-reviews.json" }, "artifacts": { "hybrid": "analysis/current/analysis-report-hybrid.json", @@ -10,9 +10,9 @@ }, "channel": "canonical", "model": null, - "published_at": "2026-05-14T00:09:20Z", + "published_at": "2026-05-14T06:06:24Z", "repo": "huggingface/transformers", "schema_version": 1, - "snapshot_id": "20260514T000044Z", + "snapshot_id": "20260514T060044Z", "variant": "hybrid" } diff --git a/snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.json b/snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.json new file mode 100644 index 0000000000000000000000000000000000000000..726d415299e2df5f0eb93a4ccd6ef5cda30ff44a --- /dev/null +++ b/snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.json @@ -0,0 +1,2205 @@ +{ + "schema_version": "1.0", + "repo": "huggingface/transformers", + "snapshot_id": "20260514T060044Z", + "generated_at": "2026-05-14T06:06:23Z", + "evidence_quality": "full", + "llm_enrichment": true, + "meta_bugs": [ + { + "cluster_id": "cluster-43979-11", + "summary": "Cluster of 1 issues and 10 PRs centered on issue #43979.", + "status": "open", + "confidence": 0.8, + "canonical_issue_number": 43979, + "canonical_pr_number": 44007, + "issue_numbers": [ + 43979 + ], + "pr_numbers": [ + 43996, + 44007, + 44013, + 44044, + 44066, + 44072, + 44085, + 44129, + 44154, + 44722 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target", + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 43996, + "right_pr_number": 44007, + "code_similarity": 0.179, + "size_similarity": 0.576, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.429, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44013, + "code_similarity": 0.122, + "size_similarity": 0.318, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.392, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44044, + "code_similarity": 0.245, + "size_similarity": 0.864, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.479, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44066, + "code_similarity": 0.225, + "size_similarity": 0.818, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.408, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44072, + "code_similarity": 0.14, + "size_similarity": 0.303, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.528, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44085, + "code_similarity": 0.216, + "size_similarity": 0.783, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.398, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44129, + "code_similarity": 0.163, + "size_similarity": 0.643, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.229, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44154, + "code_similarity": 0.153, + "size_similarity": 0.535, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.31, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44722, + "code_similarity": 0.225, + "size_similarity": 0.848, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.368, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44013, + "code_similarity": 0.19, + "size_similarity": 0.553, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.531, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44044, + "code_similarity": 0.186, + "size_similarity": 0.667, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.354, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44066, + "code_similarity": 0.188, + "size_similarity": 0.704, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.315, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44072, + "code_similarity": 0.212, + "size_similarity": 0.526, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.708, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44085, + "code_similarity": 0.195, + "size_similarity": 0.735, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.318, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44129, + "code_similarity": 0.103, + "size_similarity": 0.37, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.191, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44154, + "code_similarity": 0.238, + "size_similarity": 0.93, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.344, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44722, + "code_similarity": 0.178, + "size_similarity": 0.679, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.28, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44044, + "code_similarity": 0.126, + "size_similarity": 0.368, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.351, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44066, + "code_similarity": 0.127, + "size_similarity": 0.389, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.325, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44072, + "code_similarity": 0.29, + "size_similarity": 0.952, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.667, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44085, + "code_similarity": 0.131, + "size_similarity": 0.406, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.329, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44129, + "code_similarity": 0.07, + "size_similarity": 0.205, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.192, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44154, + "code_similarity": 0.177, + "size_similarity": 0.594, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.389, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44722, + "code_similarity": 0.118, + "size_similarity": 0.375, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.287, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44066, + "code_similarity": 0.25, + "size_similarity": 0.947, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.404, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44072, + "code_similarity": 0.136, + "size_similarity": 0.351, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.442, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44085, + "code_similarity": 0.24, + "size_similarity": 0.906, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.394, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44129, + "code_similarity": 0.147, + "size_similarity": 0.555, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.243, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44154, + "code_similarity": 0.17, + "size_similarity": 0.62, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.306, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44722, + "code_similarity": 0.257, + "size_similarity": 0.982, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.402, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44066, + "right_pr_number": 44072, + "code_similarity": 0.133, + "size_similarity": 0.37, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.393, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44066, + "right_pr_number": 44085, + "code_similarity": 0.763, + "size_similarity": 0.957, + "file_overlap": 0.5, + "area_overlap": 0.825, + "patch_similarity": 0.887, + "shared_filenames": [ + "src/transformers/models/gptj/modeling_gptj.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/models/gptj/modeling_gptj.py", + "left_ranges": [ + [ + 33, + 41 + ], + [ + 174, + 181 + ], + [ + 250, + 257 + ], + [ + 398, + 419 + ], + [ + 425, + 434 + ], + [ + 459, + 466 + ], + [ + 471, + 488 + ], + [ + 518, + 540 + ], + [ + 554, + 560 + ], + [ + 566, + 575 + ], + [ + 580, + 598 + ], + [ + 601, + 606 + ], + [ + 634, + 640 + ], + [ + 646, + 653 + ], + [ + 658, + 674 + ], + [ + 717, + 722 + ], + [ + 738, + 744 + ], + [ + 749, + 772 + ], + [ + 790, + 795 + ] + ], + "right_ranges": [ + [ + 33, + 40 + ], + [ + 173, + 180 + ], + [ + 249, + 256 + ], + [ + 397, + 405 + ], + [ + 408, + 419 + ], + [ + 425, + 434 + ], + [ + 459, + 465 + ], + [ + 470, + 489 + ], + [ + 517, + 539 + ], + [ + 553, + 559 + ], + [ + 565, + 574 + ], + [ + 579, + 597 + ], + [ + 600, + 611 + ], + [ + 633, + 639 + ], + [ + 645, + 652 + ], + [ + 657, + 673 + ], + [ + 716, + 728 + ], + [ + 737, + 743 + ], + [ + 748, + 771 + ], + [ + 789, + 794 + ] + ] + } + ] + }, + { + "left_pr_number": 44066, + "right_pr_number": 44129, + "code_similarity": 0.145, + "size_similarity": 0.526, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.263, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44066, + "right_pr_number": 44154, + "code_similarity": 0.174, + "size_similarity": 0.654, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.286, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44066, + "right_pr_number": 44722, + "code_similarity": 0.907, + "size_similarity": 0.964, + "file_overlap": 1.0, + "area_overlap": 0.808, + "patch_similarity": 0.874, + "shared_filenames": [ + "src/transformers/models/codegen/modeling_codegen.py", + "src/transformers/models/gptj/modeling_gptj.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/models/codegen/modeling_codegen.py", + "left_ranges": [ + [ + 245, + 266 + ] + ], + "right_ranges": [ + [ + 228, + 234 + ] + ] + }, + { + "filename": "src/transformers/models/gptj/modeling_gptj.py", + "left_ranges": [ + [ + 33, + 41 + ], + [ + 174, + 181 + ], + [ + 250, + 257 + ], + [ + 398, + 419 + ], + [ + 425, + 434 + ], + [ + 459, + 466 + ], + [ + 471, + 488 + ], + [ + 518, + 540 + ], + [ + 554, + 560 + ], + [ + 566, + 575 + ], + [ + 580, + 598 + ], + [ + 601, + 606 + ], + [ + 634, + 640 + ], + [ + 646, + 653 + ], + [ + 658, + 674 + ], + [ + 717, + 722 + ], + [ + 738, + 744 + ], + [ + 749, + 772 + ], + [ + 790, + 795 + ] + ], + "right_ranges": [ + [ + 33, + 42 + ], + [ + 175, + 182 + ], + [ + 251, + 258 + ], + [ + 399, + 420 + ], + [ + 426, + 435 + ], + [ + 460, + 467 + ], + [ + 472, + 489 + ], + [ + 519, + 540 + ], + [ + 554, + 560 + ], + [ + 566, + 575 + ], + [ + 580, + 598 + ], + [ + 601, + 612 + ], + [ + 634, + 640 + ], + [ + 646, + 653 + ], + [ + 658, + 663 + ], + [ + 666, + 674 + ], + [ + 717, + 722 + ], + [ + 738, + 744 + ], + [ + 749, + 772 + ], + [ + 790, + 795 + ] + ] + } + ] + }, + { + "left_pr_number": 44072, + "right_pr_number": 44085, + "code_similarity": 0.137, + "size_similarity": 0.387, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.398, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44072, + "right_pr_number": 44129, + "code_similarity": 0.074, + "size_similarity": 0.195, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.231, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44072, + "right_pr_number": 44154, + "code_similarity": 0.175, + "size_similarity": 0.566, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.414, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44072, + "right_pr_number": 44722, + "code_similarity": 0.124, + "size_similarity": 0.357, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.347, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44085, + "right_pr_number": 44129, + "code_similarity": 0.141, + "size_similarity": 0.503, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.272, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44085, + "right_pr_number": 44154, + "code_similarity": 0.18, + "size_similarity": 0.684, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.289, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44085, + "right_pr_number": 44722, + "code_similarity": 0.728, + "size_similarity": 0.923, + "file_overlap": 0.5, + "area_overlap": 0.791, + "patch_similarity": 0.78, + "shared_filenames": [ + "src/transformers/models/gptj/modeling_gptj.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/models/gptj/modeling_gptj.py", + "left_ranges": [ + [ + 33, + 40 + ], + [ + 173, + 180 + ], + [ + 249, + 256 + ], + [ + 397, + 405 + ], + [ + 408, + 419 + ], + [ + 425, + 434 + ], + [ + 459, + 465 + ], + [ + 470, + 489 + ], + [ + 517, + 539 + ], + [ + 553, + 559 + ], + [ + 565, + 574 + ], + [ + 579, + 597 + ], + [ + 600, + 611 + ], + [ + 633, + 639 + ], + [ + 645, + 652 + ], + [ + 657, + 673 + ], + [ + 716, + 728 + ], + [ + 737, + 743 + ], + [ + 748, + 771 + ], + [ + 789, + 794 + ] + ], + "right_ranges": [ + [ + 33, + 42 + ], + [ + 175, + 182 + ], + [ + 251, + 258 + ], + [ + 399, + 420 + ], + [ + 426, + 435 + ], + [ + 460, + 467 + ], + [ + 472, + 489 + ], + [ + 519, + 540 + ], + [ + 554, + 560 + ], + [ + 566, + 575 + ], + [ + 580, + 598 + ], + [ + 601, + 612 + ], + [ + 634, + 640 + ], + [ + 646, + 653 + ], + [ + 658, + 663 + ], + [ + 666, + 674 + ], + [ + 717, + 722 + ], + [ + 738, + 744 + ], + [ + 749, + 772 + ], + [ + 790, + 795 + ] + ] + } + ] + }, + { + "left_pr_number": 44129, + "right_pr_number": 44154, + "code_similarity": 0.099, + "size_similarity": 0.344, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.199, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44129, + "right_pr_number": 44722, + "code_similarity": 0.146, + "size_similarity": 0.545, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.247, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44154, + "right_pr_number": 44722, + "code_similarity": 0.164, + "size_similarity": 0.631, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.255, + "shared_filenames": [], + "shared_file_areas": [] + } + ] + }, + { + "cluster_id": "cluster-41211-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #41211.", + "status": "open", + "confidence": 0.55, + "canonical_issue_number": 41211, + "canonical_pr_number": 44339, + "issue_numbers": [ + 41211 + ], + "pr_numbers": [ + 41356, + 44339 + ], + "evidence_types": [ + "closing_reference" + ], + "pr_comparisons": [ + { + "left_pr_number": 41356, + "right_pr_number": 44339, + "code_similarity": 0.155, + "size_similarity": 0.094, + "file_overlap": 0.273, + "area_overlap": 0.078, + "patch_similarity": 0.183, + "shared_filenames": [ + "docs/source/en/model_doc/deimv2.md", + "src/transformers/models/deimv2/__init__.py", + "src/transformers/models/deimv2/configuration_deimv2.py", + "src/transformers/models/deimv2/modeling_deimv2.py", + "tests/models/deimv2/__init__.py", + "tests/models/deimv2/test_modeling_deimv2.py" + ], + "shared_file_areas": [ + { + "filename": "docs/source/en/model_doc/deimv2.md", + "left_ranges": [ + [ + 1, + 132 + ] + ], + "right_ranges": [ + [ + 1, + 65 + ] + ] + }, + { + "filename": "src/transformers/models/deimv2/__init__.py", + "left_ranges": [ + [ + 1, + 15 + ] + ], + "right_ranges": [ + [ + 1, + 29 + ] + ] + }, + { + "filename": "src/transformers/models/deimv2/configuration_deimv2.py", + "left_ranges": [ + [ + 1, + 74 + ] + ], + "right_ranges": [ + [ + 1, + 266 + ] + ] + }, + { + "filename": "tests/models/deimv2/test_modeling_deimv2.py", + "left_ranges": [ + [ + 1, + 15 + ] + ], + "right_ranges": [ + [ + 1, + 1734 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-43656-4", + "summary": "Cluster of 1 issues and 3 PRs centered on issue #43824.", + "status": "open", + "confidence": 0.8, + "canonical_issue_number": 43824, + "canonical_pr_number": 43656, + "issue_numbers": [ + 43824 + ], + "pr_numbers": [ + 43656, + 43836, + 43842 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target", + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 43656, + "right_pr_number": 43836, + "code_similarity": 0.36, + "size_similarity": 0.176, + "file_overlap": 1.0, + "area_overlap": 0.051, + "patch_similarity": 0.048, + "shared_filenames": [ + "src/transformers/cli/serve.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/cli/serve.py", + "left_ranges": [ + [ + 11, + 18 + ], + [ + 30, + 36 + ], + [ + 315, + 323 + ], + [ + 665, + 671 + ], + [ + 931, + 937 + ], + [ + 1843, + 1849 + ], + [ + 1868, + 1874 + ] + ], + "right_ranges": [ + [ + 11, + 18 + ], + [ + 359, + 429 + ], + [ + 584, + 590 + ], + [ + 1892, + 1910 + ], + [ + 1917, + 1923 + ] + ] + } + ] + }, + { + "left_pr_number": 43656, + "right_pr_number": 43842, + "code_similarity": 0.405, + "size_similarity": 0.5, + "file_overlap": 1.0, + "area_overlap": 0.0, + "patch_similarity": 0.036, + "shared_filenames": [ + "src/transformers/cli/serve.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/cli/serve.py", + "left_ranges": [ + [ + 11, + 18 + ], + [ + 30, + 36 + ], + [ + 315, + 323 + ], + [ + 665, + 671 + ], + [ + 931, + 937 + ], + [ + 1843, + 1849 + ], + [ + 1868, + 1874 + ] + ], + "right_ranges": [ + [ + 54, + 61 + ], + [ + 587, + 593 + ] + ] + } + ] + }, + { + "left_pr_number": 43836, + "right_pr_number": 43842, + "code_similarity": 0.332, + "size_similarity": 0.088, + "file_overlap": 1.0, + "area_overlap": 0.033, + "patch_similarity": 0.017, + "shared_filenames": [ + "src/transformers/cli/serve.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/cli/serve.py", + "left_ranges": [ + [ + 11, + 18 + ], + [ + 359, + 429 + ], + [ + 584, + 590 + ], + [ + 1892, + 1910 + ], + [ + 1917, + 1923 + ] + ], + "right_ranges": [ + [ + 54, + 61 + ], + [ + 587, + 593 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-43240-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #43240.", + "status": "open", + "confidence": 0.75, + "canonical_issue_number": 43240, + "canonical_pr_number": 43251, + "issue_numbers": [ + 43240 + ], + "pr_numbers": [ + 43251, + 43254 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target" + ], + "pr_comparisons": [ + { + "left_pr_number": 43251, + "right_pr_number": 43254, + "code_similarity": 0.794, + "size_similarity": 0.64, + "file_overlap": 1.0, + "area_overlap": 0.667, + "patch_similarity": 0.882, + "shared_filenames": [ + "src/transformers/loss/loss_utils.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/loss/loss_utils.py", + "left_ranges": [ + [ + 30, + 50 + ] + ], + "right_ranges": [ + [ + 30, + 43 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-41115-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #41115.", + "status": "open", + "confidence": 0.55, + "canonical_issue_number": 41115, + "canonical_pr_number": 45613, + "issue_numbers": [ + 41115 + ], + "pr_numbers": [ + 41116, + 45613 + ], + "evidence_types": [ + "closing_reference" + ], + "pr_comparisons": [ + { + "left_pr_number": 41116, + "right_pr_number": 45613, + "code_similarity": 0.805, + "size_similarity": 0.885, + "file_overlap": 0.75, + "area_overlap": 0.854, + "patch_similarity": 0.693, + "shared_filenames": [ + "docs/source/en/_toctree.yml", + "docs/source/en/model_doc/minicpm3.md", + "src/transformers/models/auto/modeling_auto.py", + "src/transformers/models/minicpm3/__init__.py", + "src/transformers/models/minicpm3/configuration_minicpm3.py", + "src/transformers/models/minicpm3/modeling_minicpm3.py", + "src/transformers/models/minicpm3/modular_minicpm3.py", + "tests/models/minicpm3/__init__.py", + "tests/models/minicpm3/test_modeling_minicpm3.py" + ], + "shared_file_areas": [ + { + "filename": "docs/source/en/_toctree.yml", + "left_ranges": [ + [ + 713, + 720 + ] + ], + "right_ranges": [ + [ + 713, + 720 + ] + ] + }, + { + "filename": "docs/source/en/model_doc/minicpm3.md", + "left_ranges": [ + [ + 1, + 67 + ] + ], + "right_ranges": [ + [ + 1, + 45 + ] + ] + }, + { + "filename": "src/transformers/models/auto/modeling_auto.py", + "left_ranges": [ + [ + 285, + 291 + ], + [ + 700, + 706 + ], + [ + 1298, + 1304 + ] + ], + "right_ranges": [ + [ + 699, + 705 + ], + [ + 1301, + 1307 + ] + ] + }, + { + "filename": "src/transformers/models/minicpm3/__init__.py", + "left_ranges": [ + [ + 1, + 29 + ] + ], + "right_ranges": [ + [ + 1, + 29 + ] + ] + }, + { + "filename": "src/transformers/models/minicpm3/configuration_minicpm3.py", + "left_ranges": [ + [ + 1, + 141 + ] + ], + "right_ranges": [ + [ + 1, + 126 + ] + ] + }, + { + "filename": "src/transformers/models/minicpm3/modeling_minicpm3.py", + "left_ranges": [ + [ + 1, + 544 + ] + ], + "right_ranges": [ + [ + 1, + 522 + ] + ] + }, + { + "filename": "src/transformers/models/minicpm3/modular_minicpm3.py", + "left_ranges": [ + [ + 1, + 444 + ] + ], + "right_ranges": [ + [ + 1, + 342 + ] + ] + }, + { + "filename": "tests/models/minicpm3/test_modeling_minicpm3.py", + "left_ranges": [ + [ + 1, + 109 + ] + ], + "right_ranges": [ + [ + 1, + 136 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-45647-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #45647.", + "status": "open", + "confidence": 0.75, + "canonical_issue_number": 45647, + "canonical_pr_number": 45738, + "issue_numbers": [ + 45647 + ], + "pr_numbers": [ + 45737, + 45738 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target" + ], + "pr_comparisons": [ + { + "left_pr_number": 45737, + "right_pr_number": 45738, + "code_similarity": 0.526, + "size_similarity": 0.11, + "file_overlap": 0.5, + "area_overlap": 1.0, + "patch_similarity": 0.024, + "shared_filenames": [ + "src/transformers/models/musicgen_melody/modeling_musicgen_melody.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/models/musicgen_melody/modeling_musicgen_melody.py", + "left_ranges": [ + [ + 491, + 497 + ] + ], + "right_ranges": [ + [ + 491, + 497 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-45081-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #45081.", + "status": "open", + "confidence": 0.8, + "canonical_issue_number": 45081, + "canonical_pr_number": 45317, + "issue_numbers": [ + 45081 + ], + "pr_numbers": [ + 45086, + 45317 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target", + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 45086, + "right_pr_number": 45317, + "code_similarity": 0.589, + "size_similarity": 0.34, + "file_overlap": 0.5, + "area_overlap": 1.0, + "patch_similarity": 0.136, + "shared_filenames": [ + "src/transformers/tokenization_utils_tokenizers.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/tokenization_utils_tokenizers.py", + "left_ranges": [ + [ + 1360, + 1370 + ], + [ + 1374, + 1380 + ] + ], + "right_ranges": [ + [ + 1360, + 1370 + ], + [ + 1374, + 1380 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-45561-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #45561.", + "status": "open", + "confidence": 0.8, + "canonical_issue_number": 45561, + "canonical_pr_number": 45645, + "issue_numbers": [ + 45561 + ], + "pr_numbers": [ + 45639, + 45645 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target", + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 45639, + "right_pr_number": 45645, + "code_similarity": 0.532, + "size_similarity": 0.619, + "file_overlap": 0.5, + "area_overlap": 0.543, + "patch_similarity": 0.451, + "shared_filenames": [ + "src/transformers/testing_utils.py", + "tests/utils/test_testing_utils.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/testing_utils.py", + "left_ranges": [ + [ + 3529, + 3536 + ], + [ + 3753, + 3787 + ] + ], + "right_ranges": [ + [ + 3525, + 3558 + ], + [ + 3782, + 3788 + ] + ] + }, + { + "filename": "tests/utils/test_testing_utils.py", + "left_ranges": [ + [ + 1, + 86 + ] + ], + "right_ranges": [ + [ + 1, + 114 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-43698-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #43698.", + "status": "open", + "confidence": 0.75, + "canonical_issue_number": 43698, + "canonical_pr_number": 43779, + "issue_numbers": [ + 43698 + ], + "pr_numbers": [ + 43779, + 43816 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target" + ], + "pr_comparisons": [ + { + "left_pr_number": 43779, + "right_pr_number": 43816, + "code_similarity": 0.418, + "size_similarity": 0.538, + "file_overlap": 1.0, + "area_overlap": 0.02, + "patch_similarity": 0.02, + "shared_filenames": [ + "src/transformers/integrations/integration_utils.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/integrations/integration_utils.py", + "left_ranges": [ + [ + 2227, + 2241 + ], + [ + 2303, + 2309 + ] + ], + "right_ranges": [ + [ + 2278, + 2291 + ], + [ + 2309, + 2322 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-44018-2", + "summary": "Cluster of 2 related pull requests linked by soft_similarity.", + "status": "open", + "confidence": 0.5, + "canonical_issue_number": null, + "canonical_pr_number": 44068, + "issue_numbers": [], + "pr_numbers": [ + 44018, + 44068 + ], + "evidence_types": [ + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 44018, + "right_pr_number": 44068, + "code_similarity": 0.766, + "size_similarity": 0.939, + "file_overlap": 1.0, + "area_overlap": 0.425, + "patch_similarity": 0.866, + "shared_filenames": [ + "src/transformers/models/gpt_neo/modeling_gpt_neo.py", + "tests/models/gpt_neo/test_modeling_gpt_neo.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/models/gpt_neo/modeling_gpt_neo.py", + "left_ranges": [ + [ + 26, + 46 + ], + [ + 138, + 143 + ], + [ + 182, + 187 + ], + [ + 283, + 295 + ], + [ + 327, + 341 + ], + [ + 348, + 354 + ], + [ + 360, + 369 + ], + [ + 399, + 406 + ], + [ + 411, + 419 + ], + [ + 428, + 433 + ], + [ + 472, + 492 + ], + [ + 507, + 513 + ], + [ + 519, + 528 + ], + [ + 541, + 559 + ], + [ + 562, + 567 + ], + [ + 595, + 601 + ], + [ + 607, + 614 + ], + [ + 627, + 643 + ], + [ + 685, + 690 + ], + [ + 708, + 714 + ], + [ + 720, + 727 + ], + [ + 740, + 757 + ], + [ + 761, + 766 + ], + [ + 780, + 786 + ], + [ + 791, + 798 + ], + [ + 807, + 822 + ], + [ + 840, + 845 + ] + ], + "right_ranges": [ + [ + 26, + 31 + ], + [ + 34, + 44 + ], + [ + 136, + 141 + ], + [ + 180, + 185 + ], + [ + 281, + 293 + ], + [ + 325, + 339 + ], + [ + 346, + 352 + ], + [ + 358, + 367 + ], + [ + 397, + 404 + ], + [ + 409, + 417 + ], + [ + 426, + 434 + ], + [ + 463, + 483 + ], + [ + 498, + 504 + ], + [ + 510, + 519 + ], + [ + 532, + 550 + ], + [ + 553, + 564 + ], + [ + 586, + 592 + ], + [ + 598, + 605 + ], + [ + 618, + 634 + ], + [ + 676, + 688 + ], + [ + 699, + 705 + ], + [ + 711, + 718 + ], + [ + 731, + 748 + ], + [ + 752, + 762 + ], + [ + 771, + 777 + ], + [ + 782, + 789 + ], + [ + 798, + 813 + ], + [ + 831, + 836 + ] + ] + }, + { + "filename": "tests/models/gpt_neo/test_modeling_gpt_neo.py", + "left_ranges": [ + [ + 458, + 464 + ] + ], + "right_ranges": [ + [ + 458, + 464 + ] + ] + } + ] + } + ] + } + ], + "duplicate_issues": [], + "duplicate_prs": [ + { + "cluster_id": "cluster-41115-3", + "canonical_pr_number": 45613, + "duplicate_pr_numbers": [ + 41116 + ], + "target_issue_number": 41115, + "reason": "PRs in cluster-41115-3 are treated as duplicates because they converge on issue #41115 with closing_reference evidence." + }, + { + "cluster_id": "cluster-41211-3", + "canonical_pr_number": 44339, + "duplicate_pr_numbers": [ + 41356 + ], + "target_issue_number": 41211, + "reason": "PRs in cluster-41211-3 are treated as duplicates because they converge on issue #41211 with closing_reference evidence." + }, + { + "cluster_id": "cluster-43240-3", + "canonical_pr_number": 43251, + "duplicate_pr_numbers": [ + 43254 + ], + "target_issue_number": 43240, + "reason": "PRs in cluster-43240-3 are treated as duplicates because they converge on issue #43240 with closing_reference, shared_issue_target evidence." + }, + { + "cluster_id": "cluster-43656-4", + "canonical_pr_number": 43656, + "duplicate_pr_numbers": [ + 43836, + 43842 + ], + "target_issue_number": 43824, + "reason": "PRs in cluster-43656-4 are treated as duplicates because they converge on issue #43824 with closing_reference, shared_issue_target, soft_similarity evidence." + }, + { + "cluster_id": "cluster-43698-3", + "canonical_pr_number": 43779, + "duplicate_pr_numbers": [ + 43816 + ], + "target_issue_number": 43698, + "reason": "PRs in cluster-43698-3 are treated as duplicates because they converge on issue #43698 with closing_reference, shared_issue_target evidence." + }, + { + "cluster_id": "cluster-43979-11", + "canonical_pr_number": 44007, + "duplicate_pr_numbers": [ + 43996, + 44013, + 44044, + 44066, + 44072, + 44085, + 44129, + 44154, + 44722 + ], + "target_issue_number": 43979, + "reason": "PRs in cluster-43979-11 are treated as duplicates because they converge on issue #43979 with closing_reference, shared_issue_target, soft_similarity evidence." + }, + { + "cluster_id": "cluster-44018-2", + "canonical_pr_number": 44068, + "duplicate_pr_numbers": [ + 44018 + ], + "target_issue_number": null, + "reason": "PRs in cluster-44018-2 are treated as duplicates because they share soft_similarity evidence." + }, + { + "cluster_id": "cluster-45081-3", + "canonical_pr_number": 45317, + "duplicate_pr_numbers": [ + 45086 + ], + "target_issue_number": 45081, + "reason": "PRs in cluster-45081-3 are treated as duplicates because they converge on issue #45081 with closing_reference, shared_issue_target, soft_similarity evidence." + }, + { + "cluster_id": "cluster-45561-3", + "canonical_pr_number": 45645, + "duplicate_pr_numbers": [ + 45639 + ], + "target_issue_number": 45561, + "reason": "PRs in cluster-45561-3 are treated as duplicates because they converge on issue #45561 with closing_reference, shared_issue_target, soft_similarity evidence." + }, + { + "cluster_id": "cluster-45647-3", + "canonical_pr_number": 45738, + "duplicate_pr_numbers": [ + 45737 + ], + "target_issue_number": 45647, + "reason": "PRs in cluster-45647-3 are treated as duplicates because they converge on issue #45647 with closing_reference, shared_issue_target evidence." + } + ], + "best_issue": { + "cluster_id": "cluster-45647-3", + "issue_number": 45647, + "reason": "Issue #45647 is the strongest global issue candidate because it is open, belongs to a cluster with 3 artifacts, and carries 12 discussion comments plus 8 inbound references.", + "score": 162.0 + }, + "best_pr": { + "cluster_id": "cluster-43979-11", + "pr_number": 44007, + "reason": "PR #44007 is the strongest global PR candidate because it is open, belongs to a cluster with 11 artifacts, links to 1 issues, and carries 2 review/discussion events.", + "score": 238.886 + } +} diff --git a/snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.llm-reviews.json b/snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.llm-reviews.json new file mode 100644 index 0000000000000000000000000000000000000000..e7cb859ca5bc72b3a515269c2c19c6fd37326284 --- /dev/null +++ b/snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.llm-reviews.json @@ -0,0 +1,19537 @@ +{ + "schema_version": "1.0", + "repo": "huggingface/transformers", + "snapshot_id": "20260514T060044Z", + "generated_at": "2026-05-14T06:06:23Z", + "model": "gpt-5.4-mini?service_tier=flex", + "reviews": [ + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:24643", + "issue:30277", + "issue:34634", + "issue:34928", + "issue:35545", + "issue:36010", + "issue:36331", + "issue:39290", + "issue:41093", + "issue:41720", + "issue:41762", + "issue:42915", + "issue:43039", + "issue:43425", + "issue:43541", + "issue:43596", + "issue:43638", + "issue:43716", + "issue:43792", + "issue:43827", + "issue:43828", + "issue:43854", + "issue:43866", + "issue:43901", + "issue:43975", + "issue:44292", + "issue:44322", + "issue:44457", + "issue:44509", + "issue:44512", + "issue:44560", + "issue:44805", + "issue:44841", + "issue:44863", + "issue:44918", + "issue:45081", + "issue:45161", + "issue:45237", + "issue:45464", + "issue:45507", + "issue:45685", + "issue:45812", + "issue:45823", + "issue:45859", + "issue:45874", + "issue:45901", + "issue:45902", + "issue:45920" + ], + "soft_pairs": [ + "issue:44322|issue:45464", + "issue:45237|issue:45685", + "issue:43716|issue:45237", + "issue:43828|issue:45237", + "issue:45237|issue:45507", + "issue:44292|issue:45237", + "issue:30277|issue:35545", + "issue:35545|issue:42915", + "issue:35545|issue:41720", + "issue:44509|issue:44512", + "issue:35545|issue:36010", + "issue:44805|issue:45823", + "issue:43638|issue:44805", + "issue:43596|issue:44805", + "issue:44805|issue:45161", + "issue:44805|issue:44918", + "issue:34634|issue:35545", + "issue:30277|issue:43638", + "issue:43596|issue:43638", + "issue:43039|issue:45161", + "issue:43975|issue:44457", + "issue:44292|issue:44457", + "issue:34928|issue:44805", + "issue:24643|issue:30277", + "issue:45859|issue:45902", + "issue:43638|issue:45161", + "issue:34928|issue:36331", + "issue:43596|issue:45161", + "issue:43716|issue:45507", + "issue:39290|issue:45081", + "issue:45823|issue:45874", + "issue:43425|issue:44292", + "issue:43828|issue:45161", + "issue:34928|issue:41720", + "issue:44918|issue:45161", + "issue:43716|issue:45685", + "issue:43541|issue:43716", + "issue:43828|issue:44292", + "issue:45507|issue:45685", + "issue:43716|issue:44292", + "issue:44292|issue:44918", + "issue:43039|issue:44918", + "issue:41762|issue:43638", + "issue:43827|issue:44512", + "issue:43716|issue:44560", + "issue:43828|issue:45685", + "issue:44560|issue:45685", + "issue:43039|issue:43828", + "issue:30277|issue:36331", + "issue:34928|issue:36010", + "issue:43039|issue:44805", + "issue:43901|issue:44512", + "issue:41762|issue:43596", + "issue:39290|issue:41720", + "issue:45237|issue:45901", + "issue:44841|issue:45823", + "issue:43039|issue:44560", + "issue:45812|issue:45920", + "issue:43866|issue:44863", + "issue:43854|issue:43866", + "issue:41093|issue:43638", + "issue:43792|issue:45901" + ], + "prepared_review_unit_hash": "49ac566543f39b872fa7fc45b0b8d1e5a252eca12eec40fe37b4bec356f92c44", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 62, + "serialized_chars": 46370, + "estimated_input_tokens": 11593, + "estimated_eval_tokens": 23442 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This is a heterogeneous cluster; most pairs are only superficially similar. The strongest duplicate groups are the ZeRO-3 Bert load failures, the Qwen3.5Moe rotary-embedding device/materialization reports, and the v5 AutoTokenizer token-ID regression reports.", + "confidence": 0.69, + "canonical_issue_reason": "issue:45920 is the broadest, most representative current report of the v5 AutoTokenizer token-ID regression across multiple model families; it subsumes the narrower Granite-specific report in issue:45812.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45920 is the best canonical issue because it describes the widest impact and the clearest reusable fix target among the accepted duplicates.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44322", + "right": "issue:45464", + "accept": false, + "reason": "Different models and different failures; not the same bug." + }, + { + "left": "issue:45237", + "right": "issue:45685", + "accept": false, + "reason": "Unrelated hardware/backend errors with no shared code-path evidence." + }, + { + "left": "issue:43716", + "right": "issue:45237", + "accept": false, + "reason": "Different model families and different failure modes." + }, + { + "left": "issue:43828", + "right": "issue:45237", + "accept": false, + "reason": "Both are dtype/device-style errors, but the model paths and bugs differ." + }, + { + "left": "issue:45237", + "right": "issue:45507", + "accept": false, + "reason": "Different model-specific failures; too broad to merge." + }, + { + "left": "issue:44292", + "right": "issue:45237", + "accept": false, + "reason": "Unrelated model/runtime issues." + }, + { + "left": "issue:30277", + "right": "issue:35545", + "accept": false, + "reason": "DeepSpeed collective mismatch vs ONNX export are unrelated." + }, + { + "left": "issue:35545", + "right": "issue:42915", + "accept": false, + "reason": "Different subsystems and different concrete failures." + }, + { + "left": "issue:35545", + "right": "issue:41720", + "accept": false, + "reason": "ONNX export and auto device mapping are unrelated." + }, + { + "left": "issue:44509", + "right": "issue:44512", + "accept": false, + "reason": "Both are docs-related, but they target different removed v5 features." + }, + { + "left": "issue:35545", + "right": "issue:36010", + "accept": false, + "reason": "ONNX export vs import error; not the same issue." + }, + { + "left": "issue:44805", + "right": "issue:45823", + "accept": false, + "reason": "Different models and different runtime paths." + }, + { + "left": "issue:43638", + "right": "issue:44805", + "accept": false, + "reason": "Same generic IndexError shape pattern, but not enough evidence of the same root cause." + }, + { + "left": "issue:43596", + "right": "issue:44805", + "accept": false, + "reason": "Same generic IndexError pattern, but distinct repros and model paths." + }, + { + "left": "issue:44805", + "right": "issue:45161", + "accept": false, + "reason": "Different models and unrelated failure classes." + }, + { + "left": "issue:43039", + "right": "issue:45161", + "accept": false, + "reason": "Different feature areas; no concrete shared bug." + }, + { + "left": "issue:43975", + "right": "issue:44457", + "accept": false, + "reason": "Tokenizer detokenization bug vs LoRA merge/save mismatch are not the same defect." + }, + { + "left": "issue:44292", + "right": "issue:44457", + "accept": false, + "reason": "Different model and different regression." + }, + { + "left": "issue:34928", + "right": "issue:44805", + "accept": false, + "reason": "Both are shape-related, but they describe different training/runtime problems." + }, + { + "left": "issue:24643", + "right": "issue:30277", + "accept": false, + "reason": "Different DeepSpeed errors and different affected code paths." + }, + { + "left": "issue:45859", + "right": "issue:45902", + "accept": true, + "reason": "Same Qwen3_5MoeTextRotaryEmbedding `inv_freq`/device-movement bug, surfaced via CPU offload vs `meta -> to_empty(cuda)` materialization." + }, + { + "left": "issue:43638", + "right": "issue:45161", + "accept": false, + "reason": "Both mention loading/runtime failures, but the model and failure mechanisms differ." + }, + { + "left": "issue:43039", + "right": "issue:43828", + "accept": false, + "reason": "Different models and different numerical mismatch symptoms." + }, + { + "left": "issue:30277", + "right": "issue:36331", + "accept": false, + "reason": "DeepSpeed collective mismatch vs trainer API signature mismatch are unrelated." + }, + { + "left": "issue:34928", + "right": "issue:36010", + "accept": false, + "reason": "Activation-checkpointing/FSDP issue vs import failure are unrelated." + }, + { + "left": "issue:43039", + "right": "issue:44805", + "accept": false, + "reason": "Different subsystems and different concrete errors." + }, + { + "left": "issue:43901", + "right": "issue:44512", + "accept": false, + "reason": "Both are docs issues, but they concern different deprecated API text." + }, + { + "left": "issue:41762", + "right": "issue:43596", + "accept": false, + "reason": "Both are ZeRO-3 loading errors, but the model contexts differ and the shared root cause is not established." + }, + { + "left": "issue:39290", + "right": "issue:41720", + "accept": false, + "reason": "Different models and different runtime failures." + }, + { + "left": "issue:45237", + "right": "issue:45901", + "accept": false, + "reason": "Unrelated model/backend failure vs table QA crash." + }, + { + "left": "issue:44841", + "right": "issue:45823", + "accept": false, + "reason": "Processor failure vs device mismatch are different bugs." + }, + { + "left": "issue:43039", + "right": "issue:44560", + "accept": false, + "reason": "Cross-entropy/Liger-Kernel issue vs video StopIteration are unrelated." + }, + { + "left": "issue:45812", + "right": "issue:45920", + "accept": true, + "reason": "Same v5 AutoTokenizer token-ID regression; issue:45920 is the broader multi-model report while issue:45812 is Granite-specific." + }, + { + "left": "issue:43866", + "right": "issue:44863", + "accept": false, + "reason": "Checkpoint corruption vs load failure are different problems." + }, + { + "left": "issue:43854", + "right": "issue:43866", + "accept": false, + "reason": "Different models and unrelated failure modes." + }, + { + "left": "issue:41093", + "right": "issue:43638", + "accept": false, + "reason": "Same generic mask/index error wording, but not enough evidence of one underlying bug." + }, + { + "left": "issue:43792", + "right": "issue:45901", + "accept": false, + "reason": "Whisper model load/runtime issue vs table QA task crash are unrelated." + }, + { + "left": "issue:41762", + "right": "issue:43638", + "accept": false, + "reason": "Same broad ZeRO-3 symptom family, but the model-specific repros are too different to merge confidently." + }, + { + "left": "issue:43827", + "right": "issue:44512", + "accept": false, + "reason": "Both are docs cleanup, but they refer to different removed v5 surfaces." + }, + { + "left": "issue:43828", + "right": "issue:45161", + "accept": false, + "reason": "Different model/runtime issues; no shared concrete bug." + }, + { + "left": "issue:44918", + "right": "issue:45161", + "accept": false, + "reason": "Different model families and unrelated behavior." + }, + { + "left": "issue:43716", + "right": "issue:45685", + "accept": false, + "reason": "Different model paths and unrelated backend errors." + }, + { + "left": "issue:44292", + "right": "issue:44918", + "accept": false, + "reason": "Different bugs with different model contexts." + }, + { + "left": "issue:43039", + "right": "issue:44918", + "accept": false, + "reason": "Liger-Kernel cross-entropy issue vs SFT embedding unpacking are unrelated." + }, + { + "left": "issue:43596", + "right": "issue:43638", + "accept": true, + "reason": "Same ZeRO-3 Bert loading failure with the identical `index 0 is out of bounds for dimension 0 with size 0` error." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "Grounded and conservative overall. The accepted duplicate pairs are well-supported by the titles, and the canonical issue choice for the v5 AutoTokenizer regression is reasonable from the packet." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 2, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:36010", + "issue:39186", + "issue:39290", + "issue:40990", + "issue:41093", + "issue:41762", + "issue:43054", + "issue:43296", + "issue:43366", + "issue:43404", + "issue:43425", + "issue:43531", + "issue:43541", + "issue:43572", + "issue:43596", + "issue:43645", + "issue:43716", + "issue:43792", + "issue:43828", + "issue:43866", + "issue:43975", + "issue:43994", + "issue:44291", + "issue:44292", + "issue:44387", + "issue:44530", + "issue:44589", + "issue:44841", + "issue:44918", + "issue:44960", + "issue:45005", + "issue:45020", + "issue:45070", + "issue:45084", + "issue:45161", + "issue:45362", + "issue:45399", + "issue:45464", + "issue:45542", + "issue:45632", + "issue:45685", + "issue:45698", + "issue:45759", + "issue:45800", + "issue:45823", + "issue:45901", + "issue:45902" + ], + "soft_pairs": [ + "issue:39290|issue:43531", + "issue:43425|issue:45070", + "issue:43596|issue:45800", + "issue:43975|issue:45901", + "issue:45362|issue:45464", + "issue:43792|issue:45070", + "issue:44292|issue:45901", + "issue:43404|issue:43866", + "issue:39290|issue:45823", + "issue:45464|issue:45685", + "issue:43541|issue:45685", + "issue:43541|issue:43828", + "issue:44841|issue:45084", + "issue:36010|issue:39290", + "issue:41093|issue:41762", + "issue:45800|issue:45823", + "issue:43296|issue:43366", + "issue:43572|issue:45070", + "issue:44387|issue:45005", + "issue:43541|issue:45161", + "issue:43828|issue:45464", + "issue:44291|issue:45070", + "issue:43366|issue:45464", + "issue:45020|issue:45800", + "issue:43531|issue:45070", + "issue:41093|issue:43596", + "issue:43541|issue:44589", + "issue:43054|issue:43994", + "issue:45542|issue:45800", + "issue:45759|issue:45902", + "issue:40990|issue:44960", + "issue:39186|issue:44918", + "issue:43716|issue:44918", + "issue:45632|issue:45698", + "issue:43404|issue:45005", + "issue:43645|issue:45399", + "issue:44530|issue:45005" + ], + "prepared_review_unit_hash": "58bcb3482a849cba68ac270f6c09d4875e4c8c24d3438ab773479384c83fa1af", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 37, + "serialized_chars": 37921, + "estimated_input_tokens": 9481, + "estimated_eval_tokens": 19218 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Mostly unrelated Transformers issues are mixed together here; the soft links are generally just broad overlaps in model family, error type, or subsystem rather than the same concrete bug. I would not merge any of the soft pairs, and I\u2019d use the broad remote_code regression as the best representative issue.", + "confidence": 0.35, + "canonical_issue_reason": "issue:45020 is the broadest, most representative open bug report in the set: it describes a general regression where recent Transformers versions break models using `remote_code`, which is a common umbrella for several loading/configuration failures.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45020 is the strongest global representative because it is general, clearly scoped, and likely to cover a wider class of user-facing breakages than the more model-specific reports.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:39290", + "right": "issue:43531", + "accept": false, + "reason": "Both mention sliding-window related config issues, but they affect different model families and symptoms; not the same concrete bug." + }, + { + "left": "issue:43425", + "right": "issue:45070", + "accept": false, + "reason": "One is a Torch version incompatibility, the other is a v5 Pydantic/config regression; too different to be duplicates." + }, + { + "left": "issue:43596", + "right": "issue:45800", + "accept": false, + "reason": "Different failure modes: DeepSpeed ZeRO-3 index error vs a Torch/Transformers compatibility issue." + }, + { + "left": "issue:43975", + "right": "issue:45901", + "accept": false, + "reason": "Completely different bugs: detokenization regression vs table-question-answering crash." + }, + { + "left": "issue:45362", + "right": "issue:45464", + "accept": false, + "reason": "Both involve Qwen3.5 and chat paths, but the model size, API path, and failure mode differ; not clearly the same bug." + }, + { + "left": "issue:43792", + "right": "issue:45070", + "accept": false, + "reason": "Whisper runtime failure is unrelated to the Pydantic field regression in Transformers v5.4.0." + }, + { + "left": "issue:44292", + "right": "issue:45901", + "accept": false, + "reason": "Different domains and code paths: Qwen NVFP4 loading vs table-question-answering crash." + }, + { + "left": "issue:43404", + "right": "issue:43866", + "accept": false, + "reason": "Different models and symptoms; tied lm_head weights vs a corrupted checkpoint report." + }, + { + "left": "issue:39290", + "right": "issue:45823", + "accept": false, + "reason": "Both are Gemma-related, but one is a missing config attribute and the other is a device mismatch at forward time." + }, + { + "left": "issue:45464", + "right": "issue:45685", + "accept": false, + "reason": "Streaming chat failure on Qwen3.5 is unrelated to the MPS histogram/int error in a MoE model." + }, + { + "left": "issue:43541", + "right": "issue:45685", + "accept": false, + "reason": "Different subsystems and symptoms: torch dynamo/grouped_mm tracing error vs MPS unsupported histogram op." + }, + { + "left": "issue:43541", + "right": "issue:43828", + "accept": false, + "reason": "Both are dtype/runtime problems in MoE models, but the model families and exact code paths differ; too broad to merge." + }, + { + "left": "issue:44841", + "right": "issue:45084", + "accept": false, + "reason": "Processor failure for Voxtral is unrelated to a template-node compilation error." + }, + { + "left": "issue:36010", + "right": "issue:39290", + "accept": false, + "reason": "Import path breakage for GenerationMixin is unrelated to the Gemma sliding_window_pattern issue." + }, + { + "left": "issue:41093", + "right": "issue:41762", + "accept": false, + "reason": "Both are index/shape errors, but one is a mask-size mismatch and the other is a zero-size index during Gemma3 loading." + }, + { + "left": "issue:45800", + "right": "issue:45823", + "accept": false, + "reason": "Version incompatibility and device placement mismatch are different bugs." + }, + { + "left": "issue:43296", + "right": "issue:43366", + "accept": false, + "reason": "Different model-support gaps: PaddleOCR-VL loading in vLLM vs GGUF gpt-oss architecture support." + }, + { + "left": "issue:43572", + "right": "issue:45070", + "accept": false, + "reason": "Missing `pad_token_idx` in StableLmConfig is a model config regression, not the same as the PretrainedConfig/Pydantic issue." + }, + { + "left": "issue:44387", + "right": "issue:45005", + "accept": false, + "reason": "Both are v5 regressions, but one is int4 memory/OOM behavior and the other is tied-weights handling on translation models." + }, + { + "left": "issue:43541", + "right": "issue:45161", + "accept": false, + "reason": "Different MoE-related problems: torch dynamo grouped_mm failure vs tensor-parallel support issue for GPT-OSS." + }, + { + "left": "issue:43828", + "right": "issue:45464", + "accept": false, + "reason": "Different model families and error paths; dtype mismatch under autocast is not the same as a streaming chat API crash." + }, + { + "left": "issue:44291", + "right": "issue:45070", + "accept": false, + "reason": "init_empty_weights constructor argument failure is unrelated to the PretrainedConfig/Pydantic regression." + }, + { + "left": "issue:43366", + "right": "issue:45464", + "accept": false, + "reason": "GGUF architecture support and Qwen3.5 streaming inference failure are distinct issues." + }, + { + "left": "issue:45020", + "right": "issue:45800", + "accept": false, + "reason": "A general remote_code regression is not the same as a Torch 2.4.1/Transformers 5.8.0 incompatibility." + }, + { + "left": "issue:43531", + "right": "issue:45070", + "accept": false, + "reason": "Sliding-window config handling and Pydantic PretrainedConfig breakage are unrelated." + }, + { + "left": "issue:41093", + "right": "issue:43596", + "accept": false, + "reason": "Both are indexing errors, but the triggering conditions and code paths differ substantially." + }, + { + "left": "issue:43541", + "right": "issue:44589", + "accept": false, + "reason": "Grouped_mm tracing failure and missing Float8 storage object are different low-level runtime issues." + }, + { + "left": "issue:43054", + "right": "issue:43994", + "accept": false, + "reason": "Both concern Siglip2 quality, but one is a text-embedding regression and the other is a generic nonsensical-output report; not enough to treat as the same bug." + }, + { + "left": "issue:45542", + "right": "issue:45800", + "accept": false, + "reason": "TensorBoard-without-TensorFlow backend detection is unrelated to Torch/Transformers version incompatibility." + }, + { + "left": "issue:45759", + "right": "issue:45902", + "accept": false, + "reason": "Different bugs: config unwrapping for composite multimodal models vs uninitialized rotary embedding memory after meta/to_empty materialization." + }, + { + "left": "issue:40990", + "right": "issue:44960", + "accept": false, + "reason": "High perplexity on gpt-oss and a GLM5 issue are not the same underlying problem." + }, + { + "left": "issue:39186", + "right": "issue:44918", + "accept": false, + "reason": "FSDP weight-shape error and TRL SFT embedding unpacking failure are different training-path bugs." + }, + { + "left": "issue:43716", + "right": "issue:44918", + "accept": false, + "reason": "Mistral image dtype mismatch is unrelated to Qwen3.5 embedding unpacking in TRL." + }, + { + "left": "issue:45632", + "right": "issue:45698", + "accept": false, + "reason": "Both involve remote/custom loading, but one is a cache-path collision and the other is loading the wrong custom module after save/load; not clearly the same bug." + }, + { + "left": "issue:43404", + "right": "issue:45005", + "accept": false, + "reason": "Both mention tied weights, but they concern different models and different manifestations of the problem." + }, + { + "left": "issue:44530", + "right": "issue:45005", + "accept": false, + "reason": "PagedAttentionCache linear_attention crash is unrelated to tied-weight handling on translation models." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 3, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:36331", + "issue:38175", + "issue:39692", + "issue:41553", + "issue:41628", + "issue:41762", + "issue:42915", + "issue:42947", + "issue:43054", + "issue:43493", + "issue:43643", + "issue:43644", + "issue:43782", + "issue:43828", + "issue:43856", + "issue:43866", + "issue:43872", + "issue:43873", + "issue:43881", + "issue:43883", + "issue:43950", + "issue:43975", + "issue:44292", + "issue:44360", + "issue:44387", + "issue:44403", + "issue:44451", + "issue:44488", + "issue:44492", + "issue:44512", + "issue:44534", + "issue:44589", + "issue:44863", + "issue:44910", + "issue:44918", + "issue:44928", + "issue:44960", + "issue:45005", + "issue:45020", + "issue:45237", + "issue:45399", + "issue:45507", + "issue:45520", + "issue:45663", + "issue:45701", + "issue:45800", + "issue:45859", + "issue:45902" + ], + "soft_pairs": [ + "issue:44360|issue:44512", + "issue:42915|issue:45399", + "issue:43873|issue:45005", + "issue:44863|issue:45005", + "issue:44492|issue:44512", + "issue:44918|issue:45507", + "issue:43975|issue:44292", + "issue:43950|issue:44534", + "issue:44292|issue:45507", + "issue:39692|issue:43054", + "issue:43644|issue:43950", + "issue:43828|issue:44928", + "issue:36331|issue:41762", + "issue:43866|issue:44403", + "issue:41628|issue:45800", + "issue:43950|issue:45902", + "issue:43872|issue:44589", + "issue:43054|issue:45701", + "issue:41553|issue:43883", + "issue:38175|issue:43054", + "issue:44910|issue:45859", + "issue:45663|issue:45902", + "issue:41553|issue:43881", + "issue:42947|issue:44387", + "issue:43054|issue:43493", + "issue:44403|issue:44451", + "issue:43643|issue:45020", + "issue:43975|issue:45520", + "issue:43975|issue:45237", + "issue:44403|issue:44488", + "issue:42947|issue:43856", + "issue:43782|issue:44960", + "issue:43881|issue:43883", + "issue:41762|issue:42915" + ], + "prepared_review_unit_hash": "3baa6d1cea22003057b6dc33ee40cfc41a44e2e36fec6c92717a7314c8805d5e", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 34, + "serialized_chars": 37482, + "estimated_input_tokens": 9371, + "estimated_eval_tokens": 18998 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Most items are distinct bugs or regressions. The clearest duplicate cluster is the Transformers v5 non-persistent-buffer corruption reports around issue 43950, with 44534 and 45902 describing the same underlying regression more specifically. The remote_code issues are related in theme but not tight enough to merge as duplicates.", + "confidence": 0.68, + "canonical_issue_reason": "Issue 43950 is the best canonical anchor: it states the underlying v5 regression clearly, is broad enough to cover the related buffer-corruption reports, and has the strongest duplicate evidence.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 43950 is the strongest overall representative because it names the core regression in a reusable way and matches the closest confirmed duplicates.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44360", + "right": "issue:44512", + "accept": false, + "reason": "Different problems: a DSA indexer design discussion versus a docs typo about the removed run command." + }, + { + "left": "issue:42915", + "right": "issue:45399", + "accept": false, + "reason": "One is a Qwen3Moe FP8 failure, the other is flash-attn2 fallback gating; different failure modes and code paths." + }, + { + "left": "issue:43873", + "right": "issue:45005", + "accept": false, + "reason": "Quantization/offloading behavior is not the same bug as tied-weights handling on translation models." + }, + { + "left": "issue:44863", + "right": "issue:45005", + "accept": false, + "reason": "NemotronH checkpoint loading failure is unrelated to the tied-weights v5 issue." + }, + { + "left": "issue:44492", + "right": "issue:44512", + "accept": false, + "reason": "A cache-strategy typo is unrelated to the docs/run-command removal problem." + }, + { + "left": "issue:44918", + "right": "issue:45507", + "accept": false, + "reason": "These are different model/runtime failures with no clear shared underlying code defect." + }, + { + "left": "issue:43975", + "right": "issue:44292", + "accept": false, + "reason": "Different concrete errors: detokenization regression versus a Qwen NVFP4 runtime failure." + }, + { + "left": "issue:43950", + "right": "issue:44534", + "accept": true, + "reason": "Both report the same v5 regression where non-persistent buffers are mishandled and end up corrupted/junk-filled." + }, + { + "left": "issue:43828", + "right": "issue:44928", + "accept": false, + "reason": "Autocast dtype mismatch and RLHF NaN explosions are different symptoms with different triggers." + }, + { + "left": "issue:36331", + "right": "issue:41762", + "accept": false, + "reason": "Trainer API keyword mismatch is unrelated to Gemma3 loading under ZeRO-3." + }, + { + "left": "issue:43866", + "right": "issue:44403", + "accept": false, + "reason": "Corrupted checkpoint data is not the same as noisy model-loading output." + }, + { + "left": "issue:41628", + "right": "issue:45800", + "accept": false, + "reason": "Import/export of AutoImageProcessor is unrelated to a torch/transformers version incompatibility claim." + }, + { + "left": "issue:43950", + "right": "issue:45902", + "accept": true, + "reason": "Both describe the same non-persistent-buffer regression; 45902 is a specific manifestation during meta/to_empty materialization." + }, + { + "left": "issue:43872", + "right": "issue:44589", + "accept": false, + "reason": "bitsandbytes init-arg incompatibility is unrelated to a Float8 storage lookup failure." + }, + { + "left": "issue:43054", + "right": "issue:45701", + "accept": false, + "reason": "Poor SigLIP2 embeddings and version-dependent tokenization changes are not the same bug." + }, + { + "left": "issue:41553", + "right": "issue:43883", + "accept": false, + "reason": "Voxtral tokenizer error messaging and Molmo tied-weights attribute failure are separate issues." + }, + { + "left": "issue:38175", + "right": "issue:43054", + "accept": false, + "reason": "Both involve SigLIP2, but zero probabilities and worse embeddings are too different to treat as one duplicate bug." + }, + { + "left": "issue:44910", + "right": "issue:45859", + "accept": false, + "reason": "Flash-Attention crash from misread 3D position_ids is unrelated to CPU-offload incompatibility in Qwen3_5Moe rotary embedding." + }, + { + "left": "issue:45663", + "right": "issue:45902", + "accept": false, + "reason": "Gemma4 FSDP2 shared_kv state rebuilding is a different mechanism from non-persistent buffer corruption." + }, + { + "left": "issue:41553", + "right": "issue:43881", + "accept": false, + "reason": "Bad AutoTokenizer error handling and glm-4v loading failure are not the same concrete defect." + }, + { + "left": "issue:42947", + "right": "issue:44387", + "accept": false, + "reason": "Gradient checkpointing ineffectiveness under LoRA is different from int4 memory-reservation/OOM behavior." + }, + { + "left": "issue:43054", + "right": "issue:43493", + "accept": false, + "reason": "SigLIP2 quality regression and HF-vs-JAX implementation discrepancy are related only broadly, not the same duplicate bug." + }, + { + "left": "issue:44403", + "right": "issue:44451", + "accept": false, + "reason": "General loading noise is not the same as a specific model load failure for ScandiBERT." + }, + { + "left": "issue:43643", + "right": "issue:45020", + "accept": false, + "reason": "The config missing-fields bug is too specific to merge with a broad 'remote_code breaks models' report." + }, + { + "left": "issue:43975", + "right": "issue:45520", + "accept": false, + "reason": "Detokenization regression and Python 3.13 flash_attn import KeyError are unrelated." + }, + { + "left": "issue:43975", + "right": "issue:45237", + "accept": false, + "reason": "Detokenization behavior and AMD GPU runtime failure are different problems." + }, + { + "left": "issue:44403", + "right": "issue:44488", + "accept": false, + "reason": "Noise on loading and a model-specific load failure are not the same underlying issue." + }, + { + "left": "issue:42947", + "right": "issue:43856", + "accept": false, + "reason": "LoRA gradient checkpointing behavior is unrelated to Qwen3 MoE memory usage." + }, + { + "left": "issue:43782", + "right": "issue:44960", + "accept": false, + "reason": "Qwen3VL weight_only loading failure is not the same as the GLM5 report." + }, + { + "left": "issue:43881", + "right": "issue:43883", + "accept": false, + "reason": "glm-4v-9b loading failure and Molmo tied-weights attribute error are different code paths." + }, + { + "left": "issue:41762", + "right": "issue:42915", + "accept": false, + "reason": "ZeRO-3 Gemma3 loading failure is unrelated to Qwen3Moe FP8 configuration failure." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet and stays conservative. The two accepted 43950 pairings are well supported by the issue titles, and the rejected pairs mostly compare clearly different failure modes or only broadly related themes. One minor omission is that 43644 also looks very close to the 43950 buffer-corruption cluster, but that does not make the stated conclusions overstated." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 4, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:24643", + "issue:30064", + "issue:36331", + "issue:41093", + "issue:41553", + "issue:42491", + "issue:42915", + "issue:42947", + "issue:43257", + "issue:43366", + "issue:43404", + "issue:43638", + "issue:43742", + "issue:43792", + "issue:43824", + "issue:43827", + "issue:43828", + "issue:43854", + "issue:43866", + "issue:43872", + "issue:43940", + "issue:43950", + "issue:44220", + "issue:44291", + "issue:44368", + "issue:44479", + "issue:44488", + "issue:44509", + "issue:44530", + "issue:44560", + "issue:44661", + "issue:44805", + "issue:44863", + "issue:44928", + "issue:44991", + "issue:45020", + "issue:45081", + "issue:45305", + "issue:45313", + "issue:45325", + "issue:45356", + "issue:45399", + "issue:45538", + "issue:45588", + "issue:45685", + "issue:45823", + "issue:45902", + "issue:45923" + ], + "soft_pairs": [ + "issue:45685|issue:45823", + "issue:42491|issue:44368", + "issue:43824|issue:43872", + "issue:45020|issue:45823", + "issue:44661|issue:45020", + "issue:44991|issue:45399", + "issue:43828|issue:45823", + "issue:43827|issue:44509", + "issue:45081|issue:45399", + "issue:45020|issue:45356", + "issue:43404|issue:43950", + "issue:43742|issue:45399", + "issue:42915|issue:45020", + "issue:43257|issue:43866", + "issue:43638|issue:45823", + "issue:43792|issue:44220", + "issue:43872|issue:44291", + "issue:43940|issue:45313", + "issue:42915|issue:43872", + "issue:41553|issue:44991", + "issue:43854|issue:44863", + "issue:36331|issue:41093", + "issue:43257|issue:45305", + "issue:42947|issue:45923", + "issue:44928|issue:45588", + "issue:41093|issue:44805", + "issue:43872|issue:45538", + "issue:41553|issue:44488", + "issue:44530|issue:44863", + "issue:44479|issue:44560", + "issue:24643|issue:30064", + "issue:45325|issue:45902", + "issue:43366|issue:43828" + ], + "prepared_review_unit_hash": "b42619d9cf91349ce6952ccd555bd4da59b9225c256f799424224dcc2af123a7", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 33, + "serialized_chars": 38066, + "estimated_input_tokens": 9517, + "estimated_eval_tokens": 19290 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Mostly unrelated issues; a few soft duplicates stand out around docs, initialization compatibility, DeepSpeed ZeRO-3 loading, mask-shape IndexErrors, and Qwen video regressions.", + "confidence": 0.67, + "canonical_issue_reason": "issue:45020 is the broadest, most representative current-regression report in the set, and it provides a reasonable anchor for comparing the more model-specific loading/tokenizer failures.", + "canonical_pr_reason": null, + "best_issue_reason": "Open, high-signal umbrella issue covering recent transformer-version regressions with real user impact; broader and more reusable as a triage anchor than the narrow per-model tickets.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45685", + "right": "issue:45823", + "accept": false, + "reason": "Different symptoms and code paths: MPS histogram/int handling vs Gemma4 PLE device placement." + }, + { + "left": "issue:42491", + "right": "issue:44368", + "accept": false, + "reason": "Both mention Qwen/LoRA, but one is a version-compatibility load failure and the other is a config warning; not the same bug." + }, + { + "left": "issue:43824", + "right": "issue:43872", + "accept": false, + "reason": "Importing a missing class and a bitsandbytes `_is_hf_initialized` constructor error are unrelated failures." + }, + { + "left": "issue:45020", + "right": "issue:45823", + "accept": false, + "reason": "Remote-code breakage is a different problem from a Gemma4 device mismatch in forward." + }, + { + "left": "issue:44661", + "right": "issue:45020", + "accept": false, + "reason": "Tokenizer-mapping guard failure is not the same as the broader remote_code regression." + }, + { + "left": "issue:44991", + "right": "issue:45399", + "accept": false, + "reason": "Tokenizer loading failure for est-roberta is unrelated to flash-attn fallback gating." + }, + { + "left": "issue:43828", + "right": "issue:45823", + "accept": false, + "reason": "Different model/runtime failures: autocast dtype mismatch vs device mismatch." + }, + { + "left": "issue:43827", + "right": "issue:44509", + "accept": true, + "reason": "Same documentation regression: both still reference pipeline tasks removed in v5." + }, + { + "left": "issue:45081", + "right": "issue:45399", + "accept": false, + "reason": "Tokenizer regex patch crash is unrelated to flash-attn fallback selection." + }, + { + "left": "issue:45020", + "right": "issue:45356", + "accept": false, + "reason": "Broad remote_code regressions and Kimi tokenizer codec handling are different root causes." + }, + { + "left": "issue:43404", + "right": "issue:43950", + "accept": false, + "reason": "Weight tying in Mistral3 and non-persistent buffer corruption are distinct from each other." + }, + { + "left": "issue:43742", + "right": "issue:45399", + "accept": false, + "reason": "MobileLLM key error is unrelated to flash-attn fallback checks." + }, + { + "left": "issue:42915", + "right": "issue:45020", + "accept": false, + "reason": "Qwen3_moe LoRA compatibility between branches is a different issue from remote_code breakage." + }, + { + "left": "issue:43257", + "right": "issue:43866", + "accept": false, + "reason": "DeepSpeed weight conversion for Qwen3 MOE and a corrupted Ovis2 checkpoint are unrelated." + }, + { + "left": "issue:43638", + "right": "issue:45823", + "accept": false, + "reason": "Zero3/Bert empty-weights IndexError is not the same as Gemma4 device placement." + }, + { + "left": "issue:43792", + "right": "issue:44220", + "accept": false, + "reason": "Whisper not running is too broad to treat as the same bug as `_torch_extract_fbank_features()` without a clearer shared traceback." + }, + { + "left": "issue:43872", + "right": "issue:44291", + "accept": true, + "reason": "Same `_is_hf_initialized` incompatibility triggered during model init/loading." + }, + { + "left": "issue:43940", + "right": "issue:45313", + "accept": true, + "reason": "Both report DeepSpeed ZeRO-3 failing to load Qwen3-family weights with missing parameters." + }, + { + "left": "issue:42915", + "right": "issue:43872", + "accept": false, + "reason": "LoRA/Qwen3_moe branch compatibility is unrelated to the bitsandbytes/init_empty_weights constructor error." + }, + { + "left": "issue:41553", + "right": "issue:44991", + "accept": false, + "reason": "A bad error message for Voxtral tokenization is not the same as an actual tokenizer load failure for a different model." + }, + { + "left": "issue:43854", + "right": "issue:44863", + "accept": false, + "reason": "Different model families and different loading failures." + }, + { + "left": "issue:36331", + "right": "issue:41093", + "accept": false, + "reason": "Unexpected trainer kwargs and mask/tensor shape mismatch are unrelated." + }, + { + "left": "issue:43257", + "right": "issue:45305", + "accept": false, + "reason": "Both involve DeepSpeed, but one is weight conversion and the other is gradient averaging under GAS." + }, + { + "left": "issue:42947", + "right": "issue:45923", + "accept": false, + "reason": "Gradient checkpointing ineffective in PEFT LoRA is not the same as a missing supports_gradient_checkpointing flag on a trust_remote_code model." + }, + { + "left": "issue:44928", + "right": "issue:45588", + "accept": false, + "reason": "NaN/SDPA fallback behavior in RLHF is unrelated to the flash_attention `s_aux=None` crash." + }, + { + "left": "issue:41093", + "right": "issue:44805", + "accept": true, + "reason": "Same mask-vs-index length mismatch IndexError, just reported with different concrete sizes." + }, + { + "left": "issue:43872", + "right": "issue:45538", + "accept": false, + "reason": "bitsandbytes init compatibility and CLIPTokenizer max length are unrelated." + }, + { + "left": "issue:41553", + "right": "issue:44488", + "accept": false, + "reason": "Voxtral error-message quality and sleng-bert loading failure are different tokenizer issues." + }, + { + "left": "issue:44530", + "right": "issue:44863", + "accept": false, + "reason": "PagedAttentionCache linear_attention crash and NemotronH checkpoint loading are unrelated." + }, + { + "left": "issue:44479", + "right": "issue:44560", + "accept": true, + "reason": "Both describe the same Qwen video-input regression in 5.3.0; 44560 is a specific manifestation." + }, + { + "left": "issue:24643", + "right": "issue:30064", + "accept": false, + "reason": "DeepSpeed training weight-shape error and void segmentation-map processing are unrelated." + }, + { + "left": "issue:45325", + "right": "issue:45902", + "accept": false, + "reason": "Qwen2.5-VL rope scaling and Qwen3.5-MoE uninitialized-memory materialization are different bugs." + }, + { + "left": "issue:43366", + "right": "issue:43828", + "accept": false, + "reason": "GGUF gpt-oss support and Phi-tiny-MoE autocast dtype mismatch are unrelated." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet and stays conservative. The accepted soft-edge pairs are plausible duplicates or near-duplicates (docs v5 pipeline references, `_is_hf_initialized` init compatibility, ZeRO-3 weight loading failures, mask-shape IndexErrors, and the Qwen video regression), while the rejected pairs mostly distinguish different symptoms or root causes appropriately. No obvious overstatement of evidence." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 5, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:36010", + "issue:38175", + "issue:39692", + "issue:41720", + "issue:42491", + "issue:42617", + "issue:42915", + "issue:43317", + "issue:43366", + "issue:43531", + "issue:43550", + "issue:43596", + "issue:43632", + "issue:43638", + "issue:43644", + "issue:43645", + "issue:43646", + "issue:43708", + "issue:43720", + "issue:43824", + "issue:43950", + "issue:43957", + "issue:43994", + "issue:44075", + "issue:44155", + "issue:44451", + "issue:44488", + "issue:44534", + "issue:44560", + "issue:44589", + "issue:44661", + "issue:44756", + "issue:44805", + "issue:44877", + "issue:44910", + "issue:44918", + "issue:44960", + "issue:44977", + "issue:45030", + "issue:45276", + "issue:45313", + "issue:45357", + "issue:45488", + "issue:45584", + "issue:45663", + "issue:45812", + "issue:45823", + "issue:45834" + ], + "soft_pairs": [ + "issue:45663|issue:45823", + "issue:44155|issue:45584", + "issue:43366|issue:43531", + "issue:43950|issue:45834", + "issue:43550|issue:44910", + "issue:43824|issue:43957", + "issue:43645|issue:43646", + "issue:43366|issue:44589", + "issue:42915|issue:43957", + "issue:44560|issue:44918", + "issue:42617|issue:43366", + "issue:36010|issue:42915", + "issue:44661|issue:45823", + "issue:38175|issue:43994", + "issue:43632|issue:45834", + "issue:44560|issue:44805", + "issue:44877|issue:45030", + "issue:43638|issue:44661", + "issue:43596|issue:44661", + "issue:44661|issue:44805", + "issue:43317|issue:44756", + "issue:43644|issue:44534", + "issue:38175|issue:39692", + "issue:44960|issue:44977", + "issue:42491|issue:43720", + "issue:45313|issue:45357", + "issue:44451|issue:44488", + "issue:41720|issue:42915", + "issue:42491|issue:45276", + "issue:43708|issue:44075", + "issue:42915|issue:44756", + "issue:45488|issue:45812" + ], + "prepared_review_unit_hash": "378e2e7e9e36f7c6fa72a81f1a77a561398974c2fa40806baf4c7cb6b9dd56cb", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 32, + "serialized_chars": 37265, + "estimated_input_tokens": 9317, + "estimated_eval_tokens": 18890 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This set is mostly a mix of unrelated closed issues, with a few clear duplicate pairs around Transformers v5 regressions and SigLIP2 output breakage. Most soft-similarity links are too broad to merge; only a small number look like the same underlying bug.", + "confidence": 0.62, + "canonical_issue_reason": "issue:43644 is the cleanest canonical anchor: it states the concrete v5 regression clearly and has an essentially exact duplicate in 44534 with the same failure mode (non-persistent buffers being filled with junk).", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43644 is the strongest global representative because it is precise, reproducible, and has the most obvious duplicate evidence among the candidates.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45663", + "right": "issue:45823", + "accept": false, + "reason": "Different Gemma4 bugs: one is FSDP2 shared_kv_states rebuilding, the other is device mismatch under auto device map." + }, + { + "left": "issue:44155", + "right": "issue:45584", + "accept": false, + "reason": "Different models and failure modes: AudioFlamingo3 batched leakage vs Whisper empty transcription after align_special_tokens." + }, + { + "left": "issue:43366", + "right": "issue:43531", + "accept": false, + "reason": "Unrelated features: GGUF gpt-oss support vs a Qwen3-MoE sliding_window issue." + }, + { + "left": "issue:43950", + "right": "issue:45834", + "accept": false, + "reason": "Different bugs: buffer corruption in from_pretrained vs a Kosmos2.5 long-OCR index error." + }, + { + "left": "issue:43550", + "right": "issue:44910", + "accept": false, + "reason": "Both are attention/compile crashes, but they affect different models and code paths." + }, + { + "left": "issue:43824", + "right": "issue:43957", + "accept": false, + "reason": "ImportError for a missing class vs meta-device model loading breakage; not the same underlying defect." + }, + { + "left": "issue:43645", + "right": "issue:43646", + "accept": true, + "reason": "Both describe the same custom-model initialization regression in Transformers v5, with the notebook case being a variant of the same failure path." + }, + { + "left": "issue:43366", + "right": "issue:44589", + "accept": false, + "reason": "Different issues: GGUF architecture support vs a float8 storage lookup/type error." + }, + { + "left": "issue:42915", + "right": "issue:43957", + "accept": false, + "reason": "Qwen3Moe FP8 failure and meta-device loading problems are distinct bugs." + }, + { + "left": "issue:44560", + "right": "issue:44918", + "accept": false, + "reason": "Different Qwen3.5-related bugs: video StopIteration vs embedding unpacking in TRL SFT." + }, + { + "left": "issue:42617", + "right": "issue:43366", + "accept": false, + "reason": "3d_parallel.py runtime failure is unrelated to GGUF gpt-oss support." + }, + { + "left": "issue:36010", + "right": "issue:42915", + "accept": false, + "reason": "ImportError for GenerationMixin is unrelated to Qwen3Moe FP8 behavior." + }, + { + "left": "issue:44661", + "right": "issue:45823", + "accept": false, + "reason": "Tokenizer-mapping failure is a different code path from Gemma4 PLE device mismatch." + }, + { + "left": "issue:38175", + "right": "issue:43994", + "accept": true, + "reason": "Both report the same SigLIP2 model producing incorrect/zero outputs with the same base checkpoint." + }, + { + "left": "issue:43632", + "right": "issue:45834", + "accept": false, + "reason": "_is_hf_initialized regression and Kosmos2.5 OCR index errors are unrelated." + }, + { + "left": "issue:44560", + "right": "issue:44805", + "accept": false, + "reason": "Different failure modes: Qwen3-vl StopIteration vs a mask/tensor shape mismatch." + }, + { + "left": "issue:44877", + "right": "issue:45030", + "accept": false, + "reason": "Both involve config validation, but the models and concrete breakages are different enough to keep separate." + }, + { + "left": "issue:43638", + "right": "issue:44661", + "accept": false, + "reason": "Deepspeed ZeRO-3 BertModel indexing and tokenizer-mapping checks are unrelated bugs." + }, + { + "left": "issue:43596", + "right": "issue:44661", + "accept": false, + "reason": "Same general area of model setup, but the root causes are different." + }, + { + "left": "issue:44661", + "right": "issue:44805", + "accept": false, + "reason": "Tokenizer mapping failure and tensor mask shape mismatch are different issues." + }, + { + "left": "issue:43317", + "right": "issue:44756", + "accept": false, + "reason": "gpu+cpu offload load failure is not the same as disabling mmap to avoid OOM on Strix Halo." + }, + { + "left": "issue:43644", + "right": "issue:44534", + "accept": true, + "reason": "Exact same non-persistent-buffer regression in Transformers v5, just phrased with slightly different wording." + }, + { + "left": "issue:38175", + "right": "issue:39692", + "accept": false, + "reason": "Both mention SigLIP2, but one is wrong outputs and the other is documentation/example/model-processor mismatch plus quantization failure." + }, + { + "left": "issue:44960", + "right": "issue:44977", + "accept": false, + "reason": "GLM5 and Qwen3.5 flash-attention failures are separate model-specific bugs." + }, + { + "left": "issue:42491", + "right": "issue:43720", + "accept": false, + "reason": "LoRA Qwen3_moe compatibility with hf5.x and BitNet packed-weight loading are unrelated." + }, + { + "left": "issue:45313", + "right": "issue:45357", + "accept": false, + "reason": "Both are Qwen3.5 issues, but one is ZeRO-3 weight loading and the other is save_pretrained visual-encoder key regression." + }, + { + "left": "issue:44451", + "right": "issue:44488", + "accept": false, + "reason": "Both are BERT-model load failures, but the available evidence is too weak to call them the same bug." + }, + { + "left": "issue:41720", + "right": "issue:42915", + "accept": false, + "reason": "Qwen3 auto device-mapping cudaErrorAssert and Qwen3Moe FP8 failures are different code paths." + }, + { + "left": "issue:42491", + "right": "issue:45276", + "accept": false, + "reason": "Qwen3_moe LoRA compatibility and Gemma4 resize_token_embeddings behavior are unrelated." + }, + { + "left": "issue:43708", + "right": "issue:44075", + "accept": false, + "reason": "Trainer max_steps resume logic and SGD args handling are separate training bugs." + }, + { + "left": "issue:42915", + "right": "issue:44756", + "accept": false, + "reason": "Qwen3Moe FP8 failure is unrelated to Strix Halo mmap OOM avoidance." + }, + { + "left": "issue:45488", + "right": "issue:45812", + "accept": false, + "reason": "Tokenizer pre-tokenizer override and Granite token-id regression are both tokenizer-related, but not clearly the same defect." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet and stays conservative overall. The strongest canonical choice, issue:43644, is well supported by the near-exact duplicate pair with issue:44534. The accepted soft pairs also look reasonably defensible from the titles alone (especially 43644/44534 and 38175/43994)." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 6, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:36010", + "issue:39404", + "issue:39692", + "issue:41720", + "issue:42491", + "issue:42947", + "issue:43257", + "issue:43335", + "issue:43381", + "issue:43404", + "issue:43425", + "issue:43454", + "issue:43644", + "issue:43645", + "issue:43854", + "issue:43856", + "issue:43873", + "issue:43931", + "issue:43957", + "issue:43994", + "issue:44155", + "issue:44164", + "issue:44291", + "issue:44292", + "issue:44387", + "issue:44402", + "issue:44488", + "issue:44743", + "issue:44756", + "issue:44910", + "issue:44912", + "issue:44928", + "issue:44960", + "issue:44991", + "issue:45216", + "issue:45276", + "issue:45292", + "issue:45335", + "issue:45362", + "issue:45406", + "issue:45588", + "issue:45663", + "issue:45684", + "issue:45698", + "issue:45704", + "issue:45759", + "issue:45823", + "issue:45859" + ], + "soft_pairs": [ + "issue:43854|issue:43931", + "issue:39692|issue:43994", + "issue:43257|issue:43854", + "issue:44488|issue:44991", + "issue:42491|issue:43931", + "issue:43644|issue:43645", + "issue:43856|issue:43873", + "issue:45362|issue:45406", + "issue:43856|issue:44387", + "issue:44910|issue:45588", + "issue:42491|issue:45292", + "issue:36010|issue:39404", + "issue:45276|issue:45335", + "issue:44164|issue:44291", + "issue:43404|issue:43454", + "issue:41720|issue:44155", + "issue:44402|issue:44488", + "issue:39692|issue:44960", + "issue:43425|issue:43957", + "issue:42947|issue:45704", + "issue:45663|issue:45759", + "issue:43335|issue:45663", + "issue:43873|issue:44387", + "issue:45663|issue:45859", + "issue:44291|issue:44756", + "issue:44155|issue:45823", + "issue:43854|issue:45216", + "issue:44292|issue:44912", + "issue:44387|issue:44756", + "issue:44743|issue:44960", + "issue:43957|issue:44756", + "issue:45684|issue:45698", + "issue:45276|issue:45292", + "issue:43381|issue:44928", + "issue:44164|issue:44756", + "issue:42947|issue:43381" + ], + "prepared_review_unit_hash": "39c67e053412b809d374ced1d9ba7be745c6ecd69d72d98ae9ef24afa7cd9090", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 36, + "serialized_chars": 38616, + "estimated_input_tokens": 9654, + "estimated_eval_tokens": 19564 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a set of unrelated Transformers issues. Only one clear duplicate pair stands out: the resize_token_embeddings regression around output embeddings, with 45276/45292/45335 describing the same underlying bug.", + "confidence": 0.81, + "canonical_issue_reason": "issue:45276 is the richest and most specific report of the clear duplicate set, covering the resize_token_embeddings/output_embeddings failure in Gemma4 and subsuming the more generic 45292/45335 wording.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45276 is the best representative issue because it names the affected model family and the concrete broken behavior, making it the most actionable canonical report.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43854", + "right": "issue:43931", + "accept": false, + "reason": "Both are model-loading failures, but they involve different models and different symptoms (unit-test loading vs weight-shape mismatch)." + }, + { + "left": "issue:39692", + "right": "issue:43994", + "accept": false, + "reason": "Both mention SigLIP2, but one is a docs/example problem and the other is a runtime output issue; not clearly the same bug." + }, + { + "left": "issue:43257", + "right": "issue:43854", + "accept": false, + "reason": "Different subsystems: Qwen3 MoE accelerate/deepspeed conversion vs GLM-4.7 load correctness." + }, + { + "left": "issue:44488", + "right": "issue:44991", + "accept": false, + "reason": "Both are tokenizer-loading regressions, but for different model repos and no evidence they share the same concrete failure." + }, + { + "left": "issue:42491", + "right": "issue:43931", + "accept": false, + "reason": "Both involve Qwen variants, but one is a LoRA save/load incompatibility and the other is a model weight-shape mismatch." + }, + { + "left": "issue:43644", + "right": "issue:43645", + "accept": false, + "reason": "Both are Transformers 5.0 regressions, but one is buffer initialization and the other is Jupyter custom-model initialization; different bugs." + }, + { + "left": "issue:43856", + "right": "issue:43873", + "accept": false, + "reason": "Both mention memory/quantization, but one is training memory efficiency and the other is offloading behavior; too broad to treat as duplicates." + }, + { + "left": "issue:45362", + "right": "issue:45406", + "accept": false, + "reason": "Different models and failure modes: chat crash for Qwen3.5 vs Gemma4Processor missing _tokenizer in serve." + }, + { + "left": "issue:43856", + "right": "issue:44387", + "accept": false, + "reason": "Related to memory, but one is MoE training efficiency and the other is int4 CUDA reserved memory/OOM; not the same underlying bug." + }, + { + "left": "issue:44910", + "right": "issue:45588", + "accept": false, + "reason": "Both are flash-attention crashes, but the stated causes are unrelated (3D position_ids packing vs s_aux=None handling)." + }, + { + "left": "issue:42491", + "right": "issue:45292", + "accept": false, + "reason": "Different issues: Qwen3 MoE LoRA compatibility vs resize_token_embeddings not updating outputs." + }, + { + "left": "issue:36010", + "right": "issue:39404", + "accept": false, + "reason": "ImportError for GenerationMixin is unrelated to Whisper pipeline return_language behavior." + }, + { + "left": "issue:45276", + "right": "issue:45335", + "accept": true, + "reason": "Both describe the same resize_token_embeddings regression where output embeddings are not updated; 45276 is just the broader Gemma4-specific phrasing." + }, + { + "left": "issue:44164", + "right": "issue:44291", + "accept": false, + "reason": "One is save/from_pretrained extra_state handling, the other is init_empty_weights argument compatibility; different code paths." + }, + { + "left": "issue:43404", + "right": "issue:43454", + "accept": false, + "reason": "Both concern untied lm_head weights, but they are model-specific reports for different architectures and are not clearly the same fix." + }, + { + "left": "issue:41720", + "right": "issue:44155", + "accept": false, + "reason": "Qwen3 device-mapping CUDA assert and AudioFlamingo3 batched inference token leakage are unrelated." + }, + { + "left": "issue:44402", + "right": "issue:44488", + "accept": false, + "reason": "Both are token/vocab-related, but one is a tokenizer vocab-size mismatch and the other is a model loading regression." + }, + { + "left": "issue:39692", + "right": "issue:44960", + "accept": false, + "reason": "SigLIP2 docs/example issues do not match the generic GLM5 loading report." + }, + { + "left": "issue:43425", + "right": "issue:43957", + "accept": false, + "reason": "Torch 2.10 incompatibility and meta-device loading breakage are different failure modes." + }, + { + "left": "issue:42947", + "right": "issue:45704", + "accept": false, + "reason": "Gradient checkpointing/PEFT LoRA is unrelated to the T5 apex.FusedRMSNorm memory leak." + }, + { + "left": "issue:45663", + "right": "issue:45759", + "accept": false, + "reason": "Gemma4 FSDP2 attention state rebuilding and Qwen multimodal config unwrapping are distinct bugs." + }, + { + "left": "issue:43335", + "right": "issue:45663", + "accept": false, + "reason": "Sparse-layer creation in SwitchTransformers and Gemma4 FSDP2 attention state handling are unrelated." + }, + { + "left": "issue:43873", + "right": "issue:44387", + "accept": false, + "reason": "Both involve quantization/memory, but offloading behavior and int4 reserved-memory growth are not the same defect." + }, + { + "left": "issue:45663", + "right": "issue:45859", + "accept": false, + "reason": "Gemma4 FSDP2 shared_kv_states handling and Qwen3_5MoeTextRotaryEmbedding CPU offload incompatibility are different issues." + }, + { + "left": "issue:44291", + "right": "issue:44756", + "accept": false, + "reason": "init_empty_weights/_is_hf_initialized and Strix Halo mmap OOM are unrelated." + }, + { + "left": "issue:44155", + "right": "issue:45823", + "accept": false, + "reason": "AudioFlamingo3 batched inference token leak does not match Gemma4 device-map mismatch." + }, + { + "left": "issue:43854", + "right": "issue:45216", + "accept": false, + "reason": "Both mention model loading, but GLM-4.7 test loading and Qwen3.5 save_pretrained correctness are different regressions." + }, + { + "left": "issue:44292", + "right": "issue:44912", + "accept": false, + "reason": "Different quantization formats and different symptoms: NVFP4 runtime error vs MXFP4 falling back to bf16." + }, + { + "left": "issue:44387", + "right": "issue:44756", + "accept": false, + "reason": "Int4 OOM under 5.x and disabling mmap on Strix Halo are unrelated memory issues." + }, + { + "left": "issue:44743", + "right": "issue:44960", + "accept": false, + "reason": "Recurrent-state reset with cache is unrelated to the GLM5 report." + }, + { + "left": "issue:43957", + "right": "issue:44756", + "accept": false, + "reason": "Meta-device loading failures and Strix Halo mmap OOM are separate problems." + }, + { + "left": "issue:45684", + "right": "issue:45698", + "accept": false, + "reason": "Both involve save/load of custom models, but one is permissions propagation and the other is loading the wrong custom module." + }, + { + "left": "issue:45276", + "right": "issue:45292", + "accept": true, + "reason": "Same underlying resize_token_embeddings bug: output embeddings are not updated; 45292 is the shorter restatement." + }, + { + "left": "issue:43381", + "right": "issue:44928", + "accept": false, + "reason": "Gradient checkpointing in eval mode and RLHF NaN explosions from 3D position_ids are different failures." + }, + { + "left": "issue:44164", + "right": "issue:44756", + "accept": false, + "reason": "save/from_pretrained extra_state handling is unrelated to mmap-related OOM on Strix Halo." + }, + { + "left": "issue:42947", + "right": "issue:43381", + "accept": false, + "reason": "Both touch gradient checkpointing, but one is PEFT LoRA ineffectiveness and the other is eval-mode prohibition; different bugs." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet and stays conservative. The only clear duplicate cluster is the resize_token_embeddings/output_embeddings regression, and the accepted soft-edge pairs are supported by the issue titles and high similarity. No overstatement stands out." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 7, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:32090", + "issue:36032", + "issue:39404", + "issue:43257", + "issue:43299", + "issue:43425", + "issue:43452", + "issue:43596", + "issue:43611", + "issue:43645", + "issue:43646", + "issue:43708", + "issue:43716", + "issue:43828", + "issue:43856", + "issue:43901", + "issue:43906", + "issue:43931", + "issue:43957", + "issue:43994", + "issue:44079", + "issue:44164", + "issue:44265", + "issue:44402", + "issue:44451", + "issue:44479", + "issue:44509", + "issue:44530", + "issue:44589", + "issue:44671", + "issue:44683", + "issue:44756", + "issue:44805", + "issue:44863", + "issue:44877", + "issue:44898", + "issue:44912", + "issue:44918", + "issue:44991", + "issue:45081", + "issue:45084", + "issue:45292", + "issue:45335", + "issue:45588", + "issue:45663", + "issue:45701", + "issue:45759", + "issue:45823" + ], + "soft_pairs": [ + "issue:44479|issue:44877", + "issue:43646|issue:43708", + "issue:45588|issue:45663", + "issue:45292|issue:45335", + "issue:43645|issue:44164", + "issue:43257|issue:44863", + "issue:44402|issue:44451", + "issue:43828|issue:43957", + "issue:43425|issue:43828", + "issue:44589|issue:44756", + "issue:44683|issue:45823", + "issue:39404|issue:45701", + "issue:43931|issue:44863", + "issue:43906|issue:44530", + "issue:43425|issue:45823", + "issue:43611|issue:43645", + "issue:43425|issue:43716", + "issue:44530|issue:45084", + "issue:43901|issue:44509", + "issue:39404|issue:44912", + "issue:44265|issue:44898", + "issue:44479|issue:44805", + "issue:43994|issue:44079", + "issue:43299|issue:43931", + "issue:39404|issue:44671", + "issue:43425|issue:44918", + "issue:44589|issue:45084", + "issue:44164|issue:45759", + "issue:36032|issue:43452", + "issue:43596|issue:44683", + "issue:44671|issue:44991", + "issue:32090|issue:39404", + "issue:43856|issue:44756", + "issue:44991|issue:45081" + ], + "prepared_review_unit_hash": "23c87f20d6abbde77d9814a668b67327c2f836c1d39af2fe181d6d0e7002d850", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 34, + "serialized_chars": 38059, + "estimated_input_tokens": 9515, + "estimated_eval_tokens": 19286 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a collection of distinct Transformers v5 regressions around model/tokenizer loading, custom initialization, and model-family-specific incompatibilities. Only one soft pair looks like the same underlying bug: resize_token_embeddings not updating dependent embeddings.", + "confidence": 0.39, + "canonical_issue_reason": "issue:43646 is the broadest, most representative user-facing regression in the set: custom model initialization breaking in Transformers 5.0. It best matches the dominant theme of loading/initialization breakages without being overly narrow.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43646 has the clearest umbrella value for triage because it captures the generic v5 initialization failure mode that several other loading-related reports orbit around.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44479", + "right": "issue:44877", + "accept": false, + "reason": "Different model families and different failure modes; both are regression reports but not the same bug." + }, + { + "left": "issue:43646", + "right": "issue:43708", + "accept": false, + "reason": "Custom model initialization breakage is unrelated to resume_from_checkpoint max_steps computation." + }, + { + "left": "issue:45588", + "right": "issue:45663", + "accept": false, + "reason": "Flash-attention s_aux=None crash and Gemma4 FSDP2 KeyError are separate code paths." + }, + { + "left": "issue:45292", + "right": "issue:45335", + "accept": true, + "reason": "Both report resize_token_embeddings failing to propagate to downstream embedding/output modules; same underlying resize logic bug." + }, + { + "left": "issue:43645", + "right": "issue:44164", + "accept": false, + "reason": "Notebook custom model init and save/from_pretrained extra_state handling are different problems." + }, + { + "left": "issue:43257", + "right": "issue:44863", + "accept": false, + "reason": "Different model-loading failures for unrelated model families and causes." + }, + { + "left": "issue:44402", + "right": "issue:44451", + "accept": false, + "reason": "Tokenizer vocab inconsistency vs a model-specific loading failure; not the same underlying defect." + }, + { + "left": "issue:43828", + "right": "issue:43957", + "accept": false, + "reason": "Autocast dtype mismatch and meta-device loading breakage are separate issues." + }, + { + "left": "issue:43425", + "right": "issue:43828", + "accept": false, + "reason": "Torch version incompatibility is not the same as a specific dtype-mismatch regression." + }, + { + "left": "issue:44589", + "right": "issue:44756", + "accept": false, + "reason": "Float8 storage lookup error and Strix Halo mmap OOM are unrelated." + }, + { + "left": "issue:44683", + "right": "issue:45823", + "accept": false, + "reason": "Compiled flex_attention failure and Gemma4 device mismatch are different bugs." + }, + { + "left": "issue:39404", + "right": "issue:45701", + "accept": false, + "reason": "Whisper return_language pipeline regression is unrelated to tokenization changes." + }, + { + "left": "issue:43931", + "right": "issue:44863", + "accept": false, + "reason": "Different checkpoint-loading failures with different model families and symptoms." + }, + { + "left": "issue:43906", + "right": "issue:44530", + "accept": false, + "reason": "An isolated reproduction report is not the same as the Qwen3.5 PagedAttentionCache crash." + }, + { + "left": "issue:43425", + "right": "issue:45823", + "accept": false, + "reason": "Torch compatibility and Gemma4 PLE device mismatch are separate issues." + }, + { + "left": "issue:43611", + "right": "issue:43645", + "accept": false, + "reason": "base_model_prefix loading regression and notebook custom init failure are related broadly, but not the same concrete bug." + }, + { + "left": "issue:43425", + "right": "issue:43716", + "accept": false, + "reason": "General torch incompatibility is not the same as the Mistral-3 image preprocessor dtype mismatch." + }, + { + "left": "issue:44530", + "right": "issue:45084", + "accept": false, + "reason": "PagedAttentionCache linear_attention crash and template compilation TypeError are unrelated." + }, + { + "left": "issue:43901", + "right": "issue:44509", + "accept": false, + "reason": "Both are docs complaints, but they cover different deprecated APIs and are not duplicates." + }, + { + "left": "issue:39404", + "right": "issue:44912", + "accept": false, + "reason": "Whisper pipeline regression and MXFP4 quantization fallback are unrelated model-path issues." + }, + { + "left": "issue:44265", + "right": "issue:44898", + "accept": false, + "reason": "torch.export/torch_compilable_check failure is unrelated to Perceiver interpolation behavior." + }, + { + "left": "issue:44479", + "right": "issue:44805", + "accept": false, + "reason": "Video input regression and mask shape mismatch are different failures." + }, + { + "left": "issue:43994", + "right": "issue:44079", + "accept": false, + "reason": "SigLIP2 incorrect outputs and ModelOutput None-key assignment are unrelated." + }, + { + "left": "issue:43299", + "right": "issue:43931", + "accept": false, + "reason": "Qwen3VL MoE load breakage and Qwen3-VL weight-shape mismatch are too different to treat as the same bug." + }, + { + "left": "issue:39404", + "right": "issue:44671", + "accept": false, + "reason": "Whisper pipeline behavior and CamemBERT masked LM predictions are unrelated." + }, + { + "left": "issue:43425", + "right": "issue:44918", + "accept": false, + "reason": "Torch incompatibility and Qwen3.5 embedding unpacking with TRL are different problems." + }, + { + "left": "issue:44589", + "right": "issue:45084", + "accept": false, + "reason": "Float8 storage lookup and non-template-node compilation errors are unrelated." + }, + { + "left": "issue:44164", + "right": "issue:45759", + "accept": false, + "reason": "extra_state save/load failure is unrelated to from_config text_config unwrapping." + }, + { + "left": "issue:36032", + "right": "issue:43452", + "accept": false, + "reason": "Tokenizer special-token method conflict and gguf_file loading regression are different bugs." + }, + { + "left": "issue:43596", + "right": "issue:44683", + "accept": false, + "reason": "Zero3/BertModel indexing error and compiled flex_attention failure are unrelated." + }, + { + "left": "issue:44671", + "right": "issue:44991", + "accept": false, + "reason": "CamemBERT MLM prediction bug and tokenizer loading failure for EMBEDDIA/est-roberta are different issues." + }, + { + "left": "issue:32090", + "right": "issue:39404", + "accept": false, + "reason": "Trainer GPU broadcast NoneType error and Whisper return_language regression are unrelated." + }, + { + "left": "issue:43856", + "right": "issue:44756", + "accept": false, + "reason": "Qwen3 MoE memory usage and Strix Halo mmap OOM are different resource issues." + }, + { + "left": "issue:44991", + "right": "issue:45081", + "accept": false, + "reason": "Tokenizer loading failure and Mistral regex patch crash are unrelated." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet and stays conservative: it characterizes the cluster as mostly distinct loading/initialization regressions without claiming duplicates beyond the resize_token_embeddings pair. The one accepted soft pair (45292/45335) is well-supported by the titles, and the other pairwise rejections are phrased cautiously." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 8, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:34567", + "issue:36032", + "issue:40444", + "issue:40990", + "issue:42915", + "issue:43262", + "issue:43317", + "issue:43322", + "issue:43377", + "issue:43404", + "issue:43421", + "issue:43452", + "issue:43482", + "issue:43540", + "issue:43541", + "issue:43582", + "issue:43611", + "issue:43632", + "issue:43643", + "issue:43644", + "issue:43645", + "issue:43646", + "issue:43716", + "issue:43883", + "issue:43994", + "issue:44164", + "issue:44206", + "issue:44220", + "issue:44451", + "issue:44466", + "issue:44479", + "issue:44488", + "issue:44530", + "issue:44534", + "issue:44610", + "issue:44661", + "issue:44683", + "issue:44737", + "issue:44811", + "issue:44849", + "issue:44987", + "issue:44991", + "issue:45081", + "issue:45084", + "issue:45325", + "issue:45507", + "issue:45584", + "issue:45859" + ], + "soft_pairs": [ + "issue:44737|issue:45859", + "issue:44164|issue:44479", + "issue:43582|issue:43632", + "issue:44488|issue:45081", + "issue:44661|issue:44683", + "issue:40990|issue:43421", + "issue:44488|issue:44987", + "issue:43611|issue:43644", + "issue:44206|issue:44220", + "issue:43716|issue:45084", + "issue:43883|issue:44488", + "issue:43643|issue:44530", + "issue:34567|issue:40444", + "issue:44811|issue:45584", + "issue:36032|issue:42915", + "issue:43644|issue:43646", + "issue:44451|issue:44987", + "issue:43452|issue:43482", + "issue:43541|issue:45084", + "issue:42915|issue:44530", + "issue:44849|issue:45325", + "issue:43317|issue:43482", + "issue:36032|issue:43645", + "issue:44987|issue:44991", + "issue:43883|issue:44991", + "issue:43404|issue:44466", + "issue:45084|issue:45507", + "issue:43262|issue:43377", + "issue:43994|issue:44610", + "issue:43646|issue:44534", + "issue:43322|issue:43540", + "issue:43645|issue:44534" + ], + "prepared_review_unit_hash": "a48c46421f835c57885b7444fe17437c40aa4cc5313d830b46738e394145001f", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 32, + "serialized_chars": 37328, + "estimated_input_tokens": 9332, + "estimated_eval_tokens": 18920 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a mix of unrelated issue reports, with only a couple of clear duplicate pockets around Transformers v5 regressions: non-persistent buffer serialization and custom model initialization. I would keep the non-persistent-buffer bug as the main canonical issue.", + "confidence": 0.74, + "canonical_issue_reason": "Issue 43644 is the most concrete representative of the duplicated v5 buffer-regression cluster, and its wording is essentially repeated by 44534.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 43644 is the best single issue to anchor the cluster because it is specific, directly duplicated, and captures the most obvious repeated failure mode.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44737", + "right": "issue:45859", + "accept": false, + "reason": "Different models and symptoms: XLNet CPU device-placement bug vs Qwen3.5 MoE CPU-offload incompatibility." + }, + { + "left": "issue:44164", + "right": "issue:44479", + "accept": false, + "reason": "Unrelated areas: save/from_pretrained extra_state handling vs video-input regression." + }, + { + "left": "issue:43582", + "right": "issue:43632", + "accept": false, + "reason": "Apple Silicon allocator TypeError and audio chat-template sampling-rate bug are unrelated." + }, + { + "left": "issue:44488", + "right": "issue:45081", + "accept": false, + "reason": "Both are tokenizer-loading complaints, but for different models and different failures; not the same concrete bug." + }, + { + "left": "issue:44661", + "right": "issue:44683", + "accept": false, + "reason": "Tokenizer mapping helper failure and compiled flex_attention regression are unrelated." + }, + { + "left": "issue:40990", + "right": "issue:43421", + "accept": false, + "reason": "Perplexity regression and runtime special-token/post-processor update bug do not share the same code-path." + }, + { + "left": "issue:44488", + "right": "issue:44987", + "accept": false, + "reason": "Both mention loading failures, but they involve different model families and different loader paths." + }, + { + "left": "issue:43611", + "right": "issue:43644", + "accept": false, + "reason": "Both are v5 regressions, but one is base_model_prefix loading and the other is non-persistent buffer serialization." + }, + { + "left": "issue:44206", + "right": "issue:44220", + "accept": false, + "reason": "Same broad audio subsystem, but the titles point to different feature-extraction failures." + }, + { + "left": "issue:43716", + "right": "issue:45084", + "accept": false, + "reason": "Mistral-3 image dtype mismatch and template-compilation error are unrelated." + }, + { + "left": "issue:43883", + "right": "issue:44488", + "accept": false, + "reason": "Different model-loading breakages with no clear shared bug." + }, + { + "left": "issue:43643", + "right": "issue:44530", + "accept": false, + "reason": "trust_remote_code config field loss and Qwen3.5 cache/group-type crash are unrelated." + }, + { + "left": "issue:34567", + "right": "issue:40444", + "accept": false, + "reason": "Trainer token accounting and multimodal iterable-dataset failure are unrelated." + }, + { + "left": "issue:44811", + "right": "issue:45584", + "accept": false, + "reason": "Both are Whisper-related, but one is batch_decode skip_special_tokens and the other is empty-transcription generation after align_special_tokens." + }, + { + "left": "issue:36032", + "right": "issue:42915", + "accept": false, + "reason": "Tokenizer method-name conflict and Qwen3Moe FP8 failure are unrelated." + }, + { + "left": "issue:43644", + "right": "issue:43646", + "accept": false, + "reason": "Non-persistent buffer junk and custom model initialization are adjacent v5 regressions, but not the same concrete bug." + }, + { + "left": "issue:44451", + "right": "issue:44987", + "accept": false, + "reason": "Both are loading regressions, but for different models and different failure modes." + }, + { + "left": "issue:43452", + "right": "issue:43482", + "accept": false, + "reason": "Both involve GGUF loading, but one is a generic from_pretrained breakage and the other is a specific Qwen2.5-GGUF v5 failure; not enough to merge strictly." + }, + { + "left": "issue:43541", + "right": "issue:45084", + "accept": false, + "reason": "Torch dynamo grouped_mm tracing failure and template compilation error are unrelated." + }, + { + "left": "issue:42915", + "right": "issue:44530", + "accept": false, + "reason": "Different Qwen3-family bugs: FP8 config failure vs linear_attention cache crash." + }, + { + "left": "issue:44849", + "right": "issue:45325", + "accept": false, + "reason": "Hidden-states bug in Qwen3.5 and Qwen2.5-VL rope-index scaling bug are different code paths." + }, + { + "left": "issue:43317", + "right": "issue:43482", + "accept": false, + "reason": "Dequantized-model device_map offload failure and Qwen2.5-GGUF loading regression are not the same bug." + }, + { + "left": "issue:36032", + "right": "issue:43645", + "accept": false, + "reason": "Tokenizer special-method conflict and custom model initialization regression are unrelated." + }, + { + "left": "issue:44987", + "right": "issue:44991", + "accept": false, + "reason": "Both are version-gated loading failures, but they affect different artifacts and loader paths." + }, + { + "left": "issue:43883", + "right": "issue:44991", + "accept": false, + "reason": "Different model-loading issues with no shared concrete root cause." + }, + { + "left": "issue:43404", + "right": "issue:44466", + "accept": false, + "reason": "Mistral3 lm_head tying bug and inconsistent tied-weight serialization are related only at a very broad theme level." + }, + { + "left": "issue:45084", + "right": "issue:45507", + "accept": false, + "reason": "Template compilation error and GraniteMoEHybrid invalid-method call are unrelated." + }, + { + "left": "issue:43262", + "right": "issue:43377", + "accept": false, + "reason": "Audio chat-template sampling-rate bug and MIMI padding-mask bug are different modalities and code paths." + }, + { + "left": "issue:43994", + "right": "issue:44610", + "accept": false, + "reason": "SigLIP2 nonsensical outputs and OmDet-Turbo image-size mismatch are unrelated." + }, + { + "left": "issue:43646", + "right": "issue:44534", + "accept": false, + "reason": "Custom model initialization regression and non-persistent buffer serialization are different failures." + }, + { + "left": "issue:43322", + "right": "issue:43540", + "accept": false, + "reason": "Llava Next segfault and Qwen3OmniMoe video ValueError are different multimodal loading/processing bugs." + }, + { + "left": "issue:43645", + "right": "issue:44534", + "accept": false, + "reason": "Notebook custom-model init breakage and non-persistent buffer serialization are not the same bug." + }, + { + "left": "issue:43645", + "right": "issue:43646", + "accept": true, + "reason": "These are the same custom-model initialization regression, with the notebook case just being a more specific manifestation." + }, + { + "left": "issue:43644", + "right": "issue:44534", + "accept": true, + "reason": "Same v5 non-persistent-buffer junk regression; the titles are essentially the same bug." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 9, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:36032", + "issue:40444", + "issue:40990", + "issue:41553", + "issue:41720", + "issue:42175", + "issue:42915", + "issue:43296", + "issue:43377", + "issue:43404", + "issue:43421", + "issue:43425", + "issue:43531", + "issue:43582", + "issue:43611", + "issue:43632", + "issue:43742", + "issue:43792", + "issue:43818", + "issue:43819", + "issue:44062", + "issue:44079", + "issue:44220", + "issue:44265", + "issue:44291", + "issue:44403", + "issue:44451", + "issue:44479", + "issue:44534", + "issue:44568", + "issue:44589", + "issue:44610", + "issue:44683", + "issue:44756", + "issue:44811", + "issue:44877", + "issue:44987", + "issue:44991", + "issue:45020", + "issue:45081", + "issue:45084", + "issue:45127", + "issue:45216", + "issue:45325", + "issue:45357", + "issue:45542", + "issue:45663" + ], + "soft_pairs": [ + "issue:44403|issue:45081", + "issue:43377|issue:43819", + "issue:42915|issue:45084", + "issue:44265|issue:44610", + "issue:44451|issue:45081", + "issue:45216|issue:45357", + "issue:44987|issue:45081", + "issue:44220|issue:44683", + "issue:43404|issue:45127", + "issue:43632|issue:44756", + "issue:42915|issue:44291", + "issue:44062|issue:44987", + "issue:44479|issue:45325", + "issue:36032|issue:43742", + "issue:43296|issue:43531", + "issue:40444|issue:45325", + "issue:42915|issue:44589", + "issue:43611|issue:44534", + "issue:43531|issue:44877", + "issue:36032|issue:41720", + "issue:43425|issue:44403", + "issue:44079|issue:45663", + "issue:40990|issue:44811", + "issue:41553|issue:43582", + "issue:43632|issue:44534", + "issue:42175|issue:45542", + "issue:43421|issue:44568", + "issue:43818|issue:45357", + "issue:43792|issue:44589", + "issue:44403|issue:44991", + "issue:44683|issue:45020" + ], + "prepared_review_unit_hash": "5c1976bcbead0059149a0f45ae54f4d13ae3c6a3ee3f3d24075ebf7ac95cae3e", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 31, + "serialized_chars": 36441, + "estimated_input_tokens": 9111, + "estimated_eval_tokens": 18478 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "These items are mostly unrelated transformer/model regressions across different subsystems. I accepted only the Qwen3.5 save_pretrained pair as the same underlying checkpoint-key regression; all other soft pairs look like separate bugs.", + "confidence": 0.81, + "canonical_issue_reason": "issue:45216 is the broadest and earliest report of the Qwen3.5 save_pretrained regression, so it is the best anchor for the only clear duplicate-like pair.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45216 is the most representative issue here because it describes the save_pretrained regression at a higher level and matches the accepted follow-up report.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44403", + "right": "issue:45081", + "accept": false, + "reason": "Different problems: generic loading noise vs Mistral tokenizer regex crash." + }, + { + "left": "issue:43377", + "right": "issue:43819", + "accept": false, + "reason": "Different code paths: MIMI padding-mask batching bug vs DAC STE mismatch." + }, + { + "left": "issue:42915", + "right": "issue:45084", + "accept": false, + "reason": "Unrelated failures: Qwen3Moe FP8 config vs template-node compilation error." + }, + { + "left": "issue:44265", + "right": "issue:44610", + "accept": false, + "reason": "Different areas: torch.export/torch_compilable_check vs OmDet processor image-size mismatch." + }, + { + "left": "issue:44451", + "right": "issue:45081", + "accept": false, + "reason": "Different models and symptoms: ScandiBERT tokenizer load failure vs Mistral tokenizer backend crash." + }, + { + "left": "issue:45216", + "right": "issue:45357", + "accept": true, + "reason": "Both report the same Qwen3.5 save_pretrained checkpoint regression, specifically incorrect saved keys/weights." + }, + { + "left": "issue:44987", + "right": "issue:45081", + "accept": false, + "reason": "Different bugs: physical-intelligence/fast loading failure vs Mistral tokenizer regex crash." + }, + { + "left": "issue:44220", + "right": "issue:44683", + "accept": false, + "reason": "Different subsystems: fbank feature extraction vs compiled flex_attention on newer torch." + }, + { + "left": "issue:43404", + "right": "issue:45127", + "accept": false, + "reason": "Different issues: untied lm_head weights vs LoRA merge collapse with extended vocab." + }, + { + "left": "issue:43632", + "right": "issue:44756", + "accept": false, + "reason": "Unrelated: _is_hf_initialized breakage vs mmap/OOM on Strix Halo." + }, + { + "left": "issue:42915", + "right": "issue:44291", + "accept": false, + "reason": "Different manifestations: Qwen3Moe FP8 config failure vs init_empty_weights argument error." + }, + { + "left": "issue:44062", + "right": "issue:44987", + "accept": false, + "reason": "Different bugs: AddedToken keyword conflict vs loading physical-intelligence/fast." + }, + { + "left": "issue:44479", + "right": "issue:45325", + "accept": false, + "reason": "Both are Qwen multimodal regressions, but they affect different code paths: video input handling vs rope index position scaling." + }, + { + "left": "issue:36032", + "right": "issue:43742", + "accept": false, + "reason": "Different models and failures: T5 tokenizer add_special_tokens conflict vs MobileLLM key error." + }, + { + "left": "issue:43296", + "right": "issue:43531", + "accept": false, + "reason": "Unrelated: PaddleOCR-VL loading in vLLM vs Qwen3-MoE sliding_window behavior." + }, + { + "left": "issue:40444", + "right": "issue:45325", + "accept": false, + "reason": "Different bugs in Qwen2.5-VL: iterable dataset multi-image failure vs still-image temporal position scaling." + }, + { + "left": "issue:42915", + "right": "issue:44589", + "accept": false, + "reason": "Different failures: FP8 config handling vs missing Float8 storage type." + }, + { + "left": "issue:43611", + "right": "issue:44534", + "accept": false, + "reason": "Different breakages: base_model_prefix loading issue vs non-persistent buffers being filled with junk." + }, + { + "left": "issue:43531", + "right": "issue:44877", + "accept": false, + "reason": "Different models and symptoms: Qwen3-MoE sliding window bug vs granite_speech config loading rejection." + }, + { + "left": "issue:36032", + "right": "issue:41720", + "accept": false, + "reason": "Completely different areas: tokenizer method conflict vs Qwen3 device-mapping CUDA assert." + }, + { + "left": "issue:43425", + "right": "issue:44403", + "accept": false, + "reason": "Unrelated: torch 2.10 compatibility vs noisy loading logs." + }, + { + "left": "issue:44079", + "right": "issue:45663", + "accept": false, + "reason": "Different problems: ModelOutput key assignment vs Gemma4 FSDP shared_kv_states KeyError." + }, + { + "left": "issue:40990", + "right": "issue:44811", + "accept": false, + "reason": "Different models and symptoms: GPT-OSS perplexity vs Whisper batch_decode skip_special_tokens bug." + }, + { + "left": "issue:41553", + "right": "issue:43582", + "accept": false, + "reason": "Unrelated: Voxtral AutoTokenizer error messaging vs Apple Silicon TypeError in warmup." + }, + { + "left": "issue:43632", + "right": "issue:44534", + "accept": false, + "reason": "Different regressions: _is_hf_initialized handling vs non-persistent buffer initialization." + }, + { + "left": "issue:42175", + "right": "issue:45542", + "accept": false, + "reason": "Related to TensorFlow packaging/backends, but not clearly the same concrete bug or mergeable fix." + }, + { + "left": "issue:43421", + "right": "issue:44568", + "accept": false, + "reason": "Both involve special tokens, but one is runtime post-processor syncing and the other is add_special_tokens not inserting BOS/EOS." + }, + { + "left": "issue:43818", + "right": "issue:45357", + "accept": false, + "reason": "Different Qwen/vision regressions: Video-LLaVA tower mismatch vs Qwen3.5 save_pretrained keys." + }, + { + "left": "issue:43792", + "right": "issue:44589", + "accept": false, + "reason": "Different failures: Whisper model runtime issue vs Float8 storage lookup error." + }, + { + "left": "issue:44403", + "right": "issue:44991", + "accept": false, + "reason": "Different loading issues: generic noisy transformer loading vs tokenizer load failure for EMBEDDIA/est-roberta." + }, + { + "left": "issue:44683", + "right": "issue:45020", + "accept": false, + "reason": "Different problems: torch>=2.9 flex_attention compilation vs remote_code model loading regressions." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "Grounded and conservative overall. The only accepted pair, issue:45216 \u2194 issue:45357, is well supported by the packet as the same Qwen3.5 `save_pretrained` checkpoint-key regression. The other soft-edge false negatives are plausible and not overstated." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 10, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:33666", + "issue:35707", + "issue:40444", + "issue:41720", + "issue:42175", + "issue:42915", + "issue:43296", + "issue:43317", + "issue:43377", + "issue:43381", + "issue:43454", + "issue:43482", + "issue:43525", + "issue:43531", + "issue:43611", + "issue:43632", + "issue:43644", + "issue:43646", + "issue:43701", + "issue:43749", + "issue:43792", + "issue:43856", + "issue:43957", + "issue:43994", + "issue:44079", + "issue:44265", + "issue:44291", + "issue:44387", + "issue:44462", + "issue:44466", + "issue:44479", + "issue:44534", + "issue:44589", + "issue:44617", + "issue:44829", + "issue:44898", + "issue:45072", + "issue:45081", + "issue:45325", + "issue:45381", + "issue:45479", + "issue:45491", + "issue:45538", + "issue:45800", + "issue:45812", + "issue:45823", + "issue:45834", + "issue:45923" + ], + "soft_pairs": [ + "issue:44462|issue:45812", + "issue:33666|issue:40444", + "issue:44829|issue:45479", + "issue:43317|issue:43856", + "issue:43482|issue:45538", + "issue:43381|issue:45923", + "issue:35707|issue:40444", + "issue:44079|issue:44534", + "issue:43482|issue:43525", + "issue:44291|issue:44589", + "issue:43377|issue:45491", + "issue:43749|issue:44079", + "issue:40444|issue:45381", + "issue:43644|issue:44079", + "issue:41720|issue:45081", + "issue:43531|issue:44291", + "issue:44265|issue:45072", + "issue:43632|issue:45072", + "issue:45325|issue:45381", + "issue:43994|issue:44898", + "issue:43611|issue:43646", + "issue:42175|issue:43296", + "issue:43454|issue:44466", + "issue:45325|issue:45834", + "issue:43792|issue:43957", + "issue:42915|issue:43482", + "issue:43296|issue:45800", + "issue:44387|issue:44617", + "issue:43701|issue:45325", + "issue:44479|issue:45823" + ], + "prepared_review_unit_hash": "46d4cd011c4b00cffd0bd48c91c8946f1cb9ec6a129c8d57f5e79d412a00318d", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 30, + "serialized_chars": 36752, + "estimated_input_tokens": 9188, + "estimated_eval_tokens": 18632 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Mostly distinct Transformers regressions across different models/subsystems. A small number of pairs look close enough to be duplicates, especially the v5 model-initialization/loading reports and the Qwen2.5-VL position-id regressions.", + "confidence": 0.64, + "canonical_issue_reason": "issue:43611 is the broadest and clearest representative of the v5 model-loading / initialization regression class, with a generic scope that can absorb closely related reports.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43611 is the best anchor issue for this cluster because it describes a concrete, widely applicable Transformers v5 breakage rather than a narrow model-specific symptom.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44462", + "right": "issue:45812", + "accept": false, + "reason": "Both are tokenizer regressions, but the symptoms differ too much: one is repository tokenizer.json lookup, the other is Granite-only token-id corruption." + }, + { + "left": "issue:33666", + "right": "issue:40444", + "accept": false, + "reason": "Different multimodal training failures: multi-GPU training vs iterable-dataset multi-image prompting." + }, + { + "left": "issue:44829", + "right": "issue:45479", + "accept": false, + "reason": "Both mention degenerate training, but one is an attention backend bug and the other is a classification-loss setup bug." + }, + { + "left": "issue:43317", + "right": "issue:43856", + "accept": false, + "reason": "Different problems: device_map/offload loading failure vs MoE training memory inefficiency." + }, + { + "left": "issue:43482", + "right": "issue:45538", + "accept": false, + "reason": "Unrelated tokenization/model-loading issue versus a CLIPTokenizer max-length constant bug." + }, + { + "left": "issue:43381", + "right": "issue:45923", + "accept": false, + "reason": "One is an eval-mode checkpointing restriction; the other is a missing model capability flag on trust_remote_code." + }, + { + "left": "issue:35707", + "right": "issue:40444", + "accept": false, + "reason": "Different generation/training scenarios and different failure modes." + }, + { + "left": "issue:44079", + "right": "issue:44534", + "accept": false, + "reason": "ModelOutput key assignment bug is unrelated to non-persistent buffer serialization junk." + }, + { + "left": "issue:43482", + "right": "issue:43525", + "accept": false, + "reason": "Both are loading-related, but one is GGUF/v5 loading and the other is a missing config attribute on Llama4." + }, + { + "left": "issue:44291", + "right": "issue:44589", + "accept": false, + "reason": "Different loading-time failures: unexpected init argument vs missing Float8 storage class." + }, + { + "left": "issue:43377", + "right": "issue:45491", + "accept": false, + "reason": "Both involve batching/padding, but they affect different models and different code paths." + }, + { + "left": "issue:43749", + "right": "issue:44079", + "accept": false, + "reason": "FSDP CPU RAM efficient loading is unrelated to ModelOutput key handling." + }, + { + "left": "issue:40444", + "right": "issue:45381", + "accept": false, + "reason": "Both concern Qwen2.5-VL/video-image inputs, but the reported bugs are different enough to keep separate." + }, + { + "left": "issue:43644", + "right": "issue:44079", + "accept": false, + "reason": "Non-persistent buffer filling is not the same issue as ModelOutput key assignment." + }, + { + "left": "issue:41720", + "right": "issue:45081", + "accept": false, + "reason": "Completely different failures: CUDA assert under device mapping vs tokenizer regex patch crash." + }, + { + "left": "issue:43531", + "right": "issue:44291", + "accept": false, + "reason": "Different regressions in Qwen3-MoE sliding-window behavior versus init_empty_weights argument handling." + }, + { + "left": "issue:44265", + "right": "issue:45072", + "accept": false, + "reason": "torch.export compatibility and dtype mismatches are different concrete bugs." + }, + { + "left": "issue:43632", + "right": "issue:45072", + "accept": false, + "reason": "_is_hf_initialized flag breakage is unrelated to bfloat16 dtype mismatches." + }, + { + "left": "issue:45325", + "right": "issue:45381", + "accept": true, + "reason": "Both point to the same Qwen2.5-VL 5.3.0 position-id / rope-index generation regression, just surfaced through still-image vs video paths." + }, + { + "left": "issue:43994", + "right": "issue:44898", + "accept": false, + "reason": "Different vision-model bugs: SigLIP2 tokenization/output issue vs Perceiver resolution/interpolation failure." + }, + { + "left": "issue:43611", + "right": "issue:43646", + "accept": true, + "reason": "Both describe Transformers 5 breaking custom-model loading/initialization on the same code path, and could plausibly be fixed together." + }, + { + "left": "issue:42175", + "right": "issue:43296", + "accept": false, + "reason": "Packaging/backend inclusion and model-loading failures are unrelated." + }, + { + "left": "issue:43454", + "right": "issue:44466", + "accept": false, + "reason": "One is tied lm_head weights causing bad generation; the other is serialization inconsistency of lm_head.weight across devices." + }, + { + "left": "issue:45325", + "right": "issue:45834", + "accept": false, + "reason": "Position-id scaling bug and Kosmos2.5 long-OCR index error are different multimodal issues." + }, + { + "left": "issue:43792", + "right": "issue:43957", + "accept": false, + "reason": "Different model-loading/runtime failures with no clear shared mechanism." + }, + { + "left": "issue:42915", + "right": "issue:43482", + "accept": false, + "reason": "Qwen3MoE FP8 config failure is unrelated to Qwen2.5 GGUF loading under v5." + }, + { + "left": "issue:43296", + "right": "issue:45800", + "accept": false, + "reason": "PaddleOCR-VL loading failure and torch/transformers version incompatibility are different root causes." + }, + { + "left": "issue:44387", + "right": "issue:44617", + "accept": false, + "reason": "Int4 reserved-memory regression and Sam3Video OOM are not the same bug." + }, + { + "left": "issue:43701", + "right": "issue:45325", + "accept": false, + "reason": "Checkpoint-resume key mismatch is unrelated to Qwen2.5-VL position-id generation." + }, + { + "left": "issue:44479", + "right": "issue:45823", + "accept": false, + "reason": "Video-input regression in Qwen-family models is unrelated to Gemma4 device mismatch during forward." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The cluster summary is grounded in the packet: issue:43611 is a reasonable broad anchor for the v5 model-loading/init regression class, and the two accepted soft edges are conservative enough based on their titles. The false soft-edge calls also stay appropriately narrow and do not overclaim duplicate status." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 11, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:34634", + "issue:35707", + "issue:39692", + "issue:40444", + "issue:40990", + "issue:41720", + "issue:42947", + "issue:43257", + "issue:43262", + "issue:43296", + "issue:43317", + "issue:43322", + "issue:43377", + "issue:43404", + "issue:43454", + "issue:43526", + "issue:43550", + "issue:43572", + "issue:43606", + "issue:43720", + "issue:43746", + "issue:43749", + "issue:43819", + "issue:43873", + "issue:43881", + "issue:43883", + "issue:44008", + "issue:44186", + "issue:44265", + "issue:44291", + "issue:44466", + "issue:44488", + "issue:44534", + "issue:44568", + "issue:44610", + "issue:44877", + "issue:44898", + "issue:44977", + "issue:44991", + "issue:45072", + "issue:45357", + "issue:45584", + "issue:45701", + "issue:45759", + "issue:45823", + "issue:45920", + "issue:45923" + ], + "soft_pairs": [ + "issue:43881|issue:44488", + "issue:43606|issue:44265", + "issue:44291|issue:44877", + "issue:44568|issue:45920", + "issue:43262|issue:43746", + "issue:40990|issue:45584", + "issue:34634|issue:35707", + "issue:43873|issue:45701", + "issue:44186|issue:44265", + "issue:35707|issue:41720", + "issue:43606|issue:43819", + "issue:41720|issue:45823", + "issue:43454|issue:43550", + "issue:43377|issue:44610", + "issue:43317|issue:43873", + "issue:39692|issue:40444", + "issue:43572|issue:45759", + "issue:43454|issue:45072", + "issue:43377|issue:44898", + "issue:43257|issue:45923", + "issue:43526|issue:43883", + "issue:40990|issue:42947", + "issue:44466|issue:45357", + "issue:43881|issue:44991", + "issue:43749|issue:44534", + "issue:43454|issue:43720", + "issue:41720|issue:43322", + "issue:43296|issue:44877", + "issue:41720|issue:44977", + "issue:44008|issue:45823", + "issue:43404|issue:45357" + ], + "prepared_review_unit_hash": "051f45a1991f839b32b3d94781da8de31ce7612c59c3f65f90893c54ae915878", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 31, + "serialized_chars": 36705, + "estimated_input_tokens": 9177, + "estimated_eval_tokens": 18610 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "These items are mostly unrelated regressions spanning loading, tokenization, multimodal processing, and training/runtime bugs. None of the soft pairs look like true duplicates; the closest recurring theme is offloading/device_map/quantization loading behavior, where issue 43873 is the broadest representative.", + "confidence": 0.63, + "canonical_issue_reason": "Issue 43873 is the broadest and most representative of the recurring offloading/quantization-loading failures in this set, with active discussion and a general symptom description.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 43873 is the best global representative because it captures the broadest underlying failure mode among the candidates and is currently open/actively relevant.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43881", + "right": "issue:44488", + "accept": false, + "reason": "Different failures in different models: glm-4v-9b load failure vs tokenizer/model loading for cjvt/sleng-bert." + }, + { + "left": "issue:43606", + "right": "issue:44265", + "accept": false, + "reason": "CPU offload device mismatch in Bark is unrelated to torch.export failures from torch_compilable_check." + }, + { + "left": "issue:44291", + "right": "issue:44877", + "accept": false, + "reason": "Unexpected _is_hf_initialized arg during init_empty_weights is a different bug from strict granite_speech config loading." + }, + { + "left": "issue:44568", + "right": "issue:45920", + "accept": false, + "reason": "One is add_special_tokens/BOS-EOS behavior for a specific tokenizer; the other is wrong token IDs across several models." + }, + { + "left": "issue:43262", + "right": "issue:43746", + "accept": false, + "reason": "Audio chat-template sampling-rate defaulting and GraniteSpeech PEFT local-checkpoint loading are separate code paths." + }, + { + "left": "issue:40990", + "right": "issue:45584", + "accept": false, + "reason": "High perplexity on gpt-oss-20b is unrelated to Whisper empty-transcription failure after align_special_tokens." + }, + { + "left": "issue:34634", + "right": "issue:35707", + "accept": false, + "reason": "Bark voice_preset not working is not the same as progressive generation with inputs_embeds and past_key_values." + }, + { + "left": "issue:43873", + "right": "issue:45701", + "accept": false, + "reason": "Offloading with quantization is unrelated to tokenizer changes across transformers versions." + }, + { + "left": "issue:44186", + "right": "issue:44265", + "accept": false, + "reason": "LayoutLMv2Tokenizer padding/NER crash is a tokenizer bug, not a torch.export/torch_compilable_check issue." + }, + { + "left": "issue:35707", + "right": "issue:41720", + "accept": false, + "reason": "Different runtime problems: generation with inputs_embeds/past_key_values versus Qwen3 device-map cudaErrorAssert." + }, + { + "left": "issue:43606", + "right": "issue:43819", + "accept": false, + "reason": "Both mention model loading behavior, but Bark CPU-offload mismatch and DAC.from_latents/STE mismatch are distinct bugs." + }, + { + "left": "issue:41720", + "right": "issue:45823", + "accept": false, + "reason": "Both involve device_map=auto, but they affect different models and failure modes (cuda assert vs forward device mismatch)." + }, + { + "left": "issue:43454", + "right": "issue:43550", + "accept": false, + "reason": "lm_head weight tying in AyaVision is unrelated to Bamba torch.compile SDPA failures." + }, + { + "left": "issue:43377", + "right": "issue:44610", + "accept": false, + "reason": "MIMI padding-mask inconsistency and OmDet-Turbo processor/model resolution mismatch are unrelated." + }, + { + "left": "issue:43317", + "right": "issue:43873", + "accept": false, + "reason": "Dequantized-model loading with gpu+cpu offload is related in theme but not the same concrete bug as offloading with quantization." + }, + { + "left": "issue:39692", + "right": "issue:40444", + "accept": false, + "reason": "SigLIP2 docs/example errors and Qwen2.5-VL iterable-dataset multi-image finetuning failures are different problems." + }, + { + "left": "issue:43572", + "right": "issue:45759", + "accept": false, + "reason": "Missing pad_token_idx in StableLmConfig is unrelated to not unwrapping text_config for multimodal Qwen configs." + }, + { + "left": "issue:43454", + "right": "issue:45072", + "accept": false, + "reason": "AyaVision tied-weights bug is unrelated to dtype mismatches in bfloat16 inference." + }, + { + "left": "issue:43377", + "right": "issue:44898", + "accept": false, + "reason": "MIMI batched-vs-single output mismatch from padding masks is unrelated to Perceiver interpolation failure." + }, + { + "left": "issue:43257", + "right": "issue:45923", + "accept": false, + "reason": "Qwen3 MOE weight conversion under accelerate+deepspeed is unrelated to a missing supports_gradient_checkpointing flag." + }, + { + "left": "issue:43526", + "right": "issue:43883", + "accept": false, + "reason": "BEiT reduce_labels returning one label is unrelated to Molmo missing all_tied_weights_keys." + }, + { + "left": "issue:40990", + "right": "issue:42947", + "accept": false, + "reason": "Perplexity regression and gradient-checkpointing inefficacy are different training/evaluation issues." + }, + { + "left": "issue:44466", + "right": "issue:45357", + "accept": false, + "reason": "lm_head serialization inconsistency by device is distinct from Qwen3.5 visual-encoder key serialization." + }, + { + "left": "issue:43881", + "right": "issue:44991", + "accept": false, + "reason": "Model loading failure for glm-4v-9b is unrelated to tokenizer loading for EMBEDDIA/est-roberta." + }, + { + "left": "issue:43749", + "right": "issue:44534", + "accept": false, + "reason": "FSDP CPU RAM efficient loading and non-persistent buffer junk are separate regressions." + }, + { + "left": "issue:43454", + "right": "issue:43720", + "accept": false, + "reason": "AyaVision weight tying and BitNet packed-weight unpacking are different loading/serialization bugs." + }, + { + "left": "issue:41720", + "right": "issue:43322", + "accept": false, + "reason": "Qwen3 device-map cuda assert and Llava Next segfault are different model-specific failures." + }, + { + "left": "issue:43296", + "right": "issue:44877", + "accept": false, + "reason": "PaddleOCR-VL load failure in vLLM is unrelated to strict granite_speech config loading." + }, + { + "left": "issue:41720", + "right": "issue:44977", + "accept": false, + "reason": "Qwen3.5 flash-attention generation bug is not the same as Qwen3 auto device mapping cuda assert." + }, + { + "left": "issue:44008", + "right": "issue:45823", + "accept": false, + "reason": "Gemma 3n variable-name collision in forward is unrelated to Gemma4 PLE device mismatch." + }, + { + "left": "issue:43404", + "right": "issue:45357", + "accept": false, + "reason": "Mistral3 lm_head tying and Qwen3.5 visual encoder key serialization are separate issues." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 12, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:36010", + "issue:36331", + "issue:40990", + "issue:41720", + "issue:42617", + "issue:42915", + "issue:43295", + "issue:43316", + "issue:43317", + "issue:43322", + "issue:43421", + "issue:43450", + "issue:43526", + "issue:43531", + "issue:43540", + "issue:43550", + "issue:43572", + "issue:43577", + "issue:43606", + "issue:43708", + "issue:43720", + "issue:43782", + "issue:43873", + "issue:43881", + "issue:44186", + "issue:44361", + "issue:44387", + "issue:44423", + "issue:44479", + "issue:44610", + "issue:44671", + "issue:44743", + "issue:44756", + "issue:44877", + "issue:44898", + "issue:44913", + "issue:44918", + "issue:44991", + "issue:45072", + "issue:45081", + "issue:45161", + "issue:45406", + "issue:45632", + "issue:45684", + "issue:45701", + "issue:45704", + "issue:45812" + ], + "soft_pairs": [ + "issue:45632|issue:45684", + "issue:42617|issue:42915", + "issue:40990|issue:45701", + "issue:44186|issue:45072", + "issue:44423|issue:45406", + "issue:43577|issue:43708", + "issue:44387|issue:45704", + "issue:43873|issue:44756", + "issue:44877|issue:44991", + "issue:36331|issue:42617", + "issue:43550|issue:43720", + "issue:36331|issue:42915", + "issue:43450|issue:45701", + "issue:44610|issue:44898", + "issue:44186|issue:44610", + "issue:44479|issue:44918", + "issue:43550|issue:43606", + "issue:41720|issue:42617", + "issue:43316|issue:44913", + "issue:42915|issue:43317", + "issue:43295|issue:45701", + "issue:43526|issue:43881", + "issue:44671|issue:44743", + "issue:41720|issue:45161", + "issue:45081|issue:45812", + "issue:43531|issue:43572", + "issue:43540|issue:43782", + "issue:36010|issue:42617", + "issue:43421|issue:45812", + "issue:44186|issue:44361", + "issue:42915|issue:43322" + ], + "prepared_review_unit_hash": "1d9946bb5e5a04a38c68fc0fa4f4f71823399f10b4fb9e95eb04324b7718722b", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 31, + "serialized_chars": 36462, + "estimated_input_tokens": 9116, + "estimated_eval_tokens": 18488 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly heterogeneous: the soft pairs share broad surface themes like tokenizers, loading, or offloading, but they do not look like the same underlying bug. I rejected all soft edges and picked the most connected issue only as a weak triage anchor, not as a true duplicate canonical.", + "confidence": 0.47, + "canonical_issue_reason": "issue:42617 is the most connected node in the soft-link graph, so it is the best weak anchor for triage, but its neighbors point to different failure modes and it is not a true duplicate center.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:42617 is the most central issue among this noisy set; it is the best representative for manual review, though not a safe duplicate target.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45632", + "right": "issue:45684", + "accept": false, + "reason": "Cache-path collisions for trust_remote_code and read-only permission propagation are different bugs." + }, + { + "left": "issue:42617", + "right": "issue:42915", + "accept": false, + "reason": "3d_parallel.py execution failure and Qwen3Moe FP8 loading failure are unrelated code paths." + }, + { + "left": "issue:40990", + "right": "issue:45701", + "accept": false, + "reason": "High perplexity on WikiText-2 and a tokenization regression are not the same failure." + }, + { + "left": "issue:44186", + "right": "issue:45072", + "accept": false, + "reason": "LayoutLMv2 tokenizer crashes and bf16 dtype mismatches have different root causes." + }, + { + "left": "issue:44423", + "right": "issue:45406", + "accept": false, + "reason": "A continuous-batching multimodal crash with str.to and a missing _tokenizer on Gemma4Processor are distinct bugs." + }, + { + "left": "issue:43577", + "right": "issue:43708", + "accept": false, + "reason": "Blip2 dtype handling and Trainer resume max_steps calculation are unrelated." + }, + { + "left": "issue:44387", + "right": "issue:45704", + "accept": false, + "reason": "Int4 quantization memory growth and an apex RMSNorm memory leak are different issues." + }, + { + "left": "issue:43873", + "right": "issue:44756", + "accept": false, + "reason": "Generic quantization offloading problems and Strix Halo mmap OOM are not the same bug." + }, + { + "left": "issue:44877", + "right": "issue:44991", + "accept": false, + "reason": "Strict config loading and tokenizer loading failure are separate code paths." + }, + { + "left": "issue:36331", + "right": "issue:42617", + "accept": false, + "reason": "CustomTrainer.compute_loss signature breakage is unrelated to 3d_parallel.py startup failure." + }, + { + "left": "issue:43550", + "right": "issue:43720", + "accept": false, + "reason": "torch.compile SDPA failure in Bamba and packed-weight unpacking in BitNet are different." + }, + { + "left": "issue:36331", + "right": "issue:42915", + "accept": false, + "reason": "Trainer compute_loss API breakage and Qwen3Moe FP8 loading failure do not match." + }, + { + "left": "issue:43450", + "right": "issue:45701", + "accept": false, + "reason": "Batched video shape handling and version-dependent tokenization changes are not the same bug." + }, + { + "left": "issue:44610", + "right": "issue:44898", + "accept": false, + "reason": "OmDet-Turbo input-size mismatch and Perceiver non-default-resolution failure are similar only at a high level, not the same code path." + }, + { + "left": "issue:44186", + "right": "issue:44610", + "accept": false, + "reason": "Tokenizer crashes in LayoutLMv2 and image-size mismatch in OmDet-Turbo are unrelated." + }, + { + "left": "issue:44479", + "right": "issue:44918", + "accept": false, + "reason": "Qwen video-input regression and Qwen3.5 SFT embedding unpacking are different failures." + }, + { + "left": "issue:43550", + "right": "issue:43606", + "accept": false, + "reason": "SDPA/compile failure and CPU-offload device mismatch are separate issues." + }, + { + "left": "issue:41720", + "right": "issue:42617", + "accept": false, + "reason": "Qwen3 auto device mapping cudaErrorAssert and 3d_parallel.py execution failure are unrelated." + }, + { + "left": "issue:43316", + "right": "issue:44913", + "accept": false, + "reason": "Gemma3TextConfig API mismatch and GPTNeoX rotary_pct reload bug are different config problems." + }, + { + "left": "issue:42915", + "right": "issue:43317", + "accept": false, + "reason": "Qwen3Moe FP8 loading and dequantized model offload failures are different quantization/loading paths." + }, + { + "left": "issue:43295", + "right": "issue:45701", + "accept": false, + "reason": "Processor.tokenizer regression and tokenization changes are not the same underlying bug." + }, + { + "left": "issue:43526", + "right": "issue:43881", + "accept": false, + "reason": "BeitImageProcessorFast label reduction bug and glm-4v-9b loading failure are unrelated." + }, + { + "left": "issue:44671", + "right": "issue:44743", + "accept": false, + "reason": "CamemBERT masked-LM regression and Qwen3_5 recurrent-state reset are distinct model bugs." + }, + { + "left": "issue:41720", + "right": "issue:45161", + "accept": false, + "reason": "Auto device mapping failure and GPT-OSS tensor-parallel failure are different model-parallelism issues." + }, + { + "left": "issue:45081", + "right": "issue:45812", + "accept": false, + "reason": "Mistral regex patch crash and Granite token-id regression are both tokenizer bugs but not the same fix." + }, + { + "left": "issue:43531", + "right": "issue:43572", + "accept": false, + "reason": "Qwen3-MoE sliding-window behavior and StableLm pad_token_idx omission are unrelated." + }, + { + "left": "issue:43540", + "right": "issue:43782", + "accept": false, + "reason": "Video-processing ValueError in Qwen3OmniMoe and weight_only loading error in Qwen3VL are different multimodal failures." + }, + { + "left": "issue:36010", + "right": "issue:42617", + "accept": false, + "reason": "GenerationMixin import error and 3d_parallel.py execution failure have no common code path." + }, + { + "left": "issue:43421", + "right": "issue:45812", + "accept": false, + "reason": "Runtime post-processor updates for special tokens and Granite token-id regression are different tokenizer behaviors." + }, + { + "left": "issue:44186", + "right": "issue:44361", + "accept": false, + "reason": "LayoutLMv2Tokenizer NER/padding crash and MLukeTokenizer task AttributeError are separate tokenizer bugs." + }, + { + "left": "issue:42915", + "right": "issue:43322", + "accept": false, + "reason": "Qwen3Moe FP8 failure and Llava Next segfault are unrelated model-loading problems." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 13, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:30064", + "issue:33357", + "issue:34567", + "issue:34634", + "issue:35707", + "issue:41720", + "issue:42490", + "issue:43329", + "issue:43388", + "issue:43421", + "issue:43531", + "issue:43550", + "issue:43606", + "issue:43653", + "issue:43720", + "issue:43746", + "issue:43782", + "issue:43873", + "issue:43927", + "issue:44062", + "issue:44077", + "issue:44368", + "issue:44451", + "issue:44492", + "issue:44509", + "issue:44514", + "issue:44545", + "issue:44589", + "issue:44625", + "issue:44898", + "issue:44918", + "issue:44936", + "issue:44987", + "issue:44991", + "issue:45042", + "issue:45059", + "issue:45072", + "issue:45357", + "issue:45381", + "issue:45406", + "issue:45464", + "issue:45479", + "issue:45701", + "issue:45704", + "issue:45759", + "issue:45859", + "issue:45923" + ], + "soft_pairs": [ + "issue:43388|issue:45059", + "issue:45406|issue:45464", + "issue:44492|issue:44509", + "issue:33357|issue:35707", + "issue:43421|issue:45701", + "issue:43606|issue:43720", + "issue:41720|issue:43746", + "issue:30064|issue:34634", + "issue:43550|issue:43653", + "issue:43873|issue:45704", + "issue:44514|issue:44545", + "issue:44625|issue:45479", + "issue:43606|issue:43653", + "issue:43531|issue:45759", + "issue:42490|issue:43329", + "issue:34634|issue:41720", + "issue:34567|issue:35707", + "issue:43421|issue:43927", + "issue:44368|issue:45357", + "issue:43782|issue:45381", + "issue:44898|issue:45072", + "issue:44451|issue:44991", + "issue:44077|issue:44625", + "issue:44987|issue:45042", + "issue:45759|issue:45859", + "issue:43531|issue:44589", + "issue:44062|issue:44451", + "issue:45704|issue:45923", + "issue:44918|issue:44936" + ], + "prepared_review_unit_hash": "ab08afc26cf4eb06793ff123305a7fe25bd97bc0b7807091c4826bf10e667441", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 29, + "serialized_chars": 36080, + "estimated_input_tokens": 9020, + "estimated_eval_tokens": 18296 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is mostly heterogeneous. Only one soft pair is a clear duplicate: the Qwen2_5_VLProcessor.apply_chat_template crash on batched input with padding=False.", + "confidence": 0.94, + "canonical_issue_reason": "Issue 44545 is the best canonical issue because it is an exact duplicate of 44514, has the clearer later report, and directly names the same Qwen2_5_VLProcessor padding=False crash.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 44545 is the strongest issue candidate overall in this cluster because it is the only clear duplicate target and most precisely represents the shared bug in the accepted pair.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43388", + "right": "issue:45059", + "accept": false, + "reason": "Different problems and code paths: metric gathering drops tuple labels vs SAM3 text/bbox behavior." + }, + { + "left": "issue:45406", + "right": "issue:45464", + "accept": false, + "reason": "Different model/runtime failures: Gemma4Processor missing _tokenizer vs Qwen3.5 chat streaming API failure." + }, + { + "left": "issue:44492", + "right": "issue:44509", + "accept": false, + "reason": "One is a typo in cache strategy docs, the other is stale pipeline docs; related documentation cleanup but not the same bug." + }, + { + "left": "issue:33357", + "right": "issue:35707", + "accept": false, + "reason": "Unrelated failures: MacOS bus error with a CLIP model vs progressive generation with inputs_embeds/past_key_values." + }, + { + "left": "issue:43421", + "right": "issue:45701", + "accept": false, + "reason": "Runtime special-token post-processor update bug is different from a broad tokenization regression report." + }, + { + "left": "issue:43606", + "right": "issue:43720", + "accept": false, + "reason": "Both involve loading/offload, but they are different concrete failures in different models: CPU offload device mismatch vs packed-weight unpacking." + }, + { + "left": "issue:41720", + "right": "issue:43746", + "accept": false, + "reason": "Different subsystems and failures: Qwen3 auto device mapping cuda assert vs GraniteSpeech PEFT local checkpoint loading." + }, + { + "left": "issue:30064", + "right": "issue:34634", + "accept": false, + "reason": "Different processors and failures: void segmentation maps vs Bark voice_preset behavior." + }, + { + "left": "issue:43550", + "right": "issue:43653", + "accept": false, + "reason": "Different areas entirely: torch.compile/SDPA issue vs tokenizer special-token registration." + }, + { + "left": "issue:43873", + "right": "issue:45704", + "accept": false, + "reason": "Quantization/offloading behavior is unrelated to a T5 apex FusedRMSNorm memory leak." + }, + { + "left": "issue:44514", + "right": "issue:44545", + "accept": true, + "reason": "Same exact Qwen2_5_VLProcessor.apply_chat_template batched-input crash with padding=False; clearly the same underlying bug." + }, + { + "left": "issue:44625", + "right": "issue:45479", + "accept": false, + "reason": "Both mention classification settings, but one is config propagation and the other is degenerate loss computation." + }, + { + "left": "issue:43606", + "right": "issue:43653", + "accept": false, + "reason": "Different failure modes and models; one is bark-small CPU offload mismatch, the other is a tokenizer special-token bug." + }, + { + "left": "issue:43531", + "right": "issue:45759", + "accept": false, + "reason": "Qwen3-MoE sliding_window behavior is not the same as unwrapping text_config for composite multimodal configs." + }, + { + "left": "issue:42490", + "right": "issue:43329", + "accept": false, + "reason": "A trailing-slash import warning is unrelated to undefined helpers in multimodal token counting." + }, + { + "left": "issue:34634", + "right": "issue:41720", + "accept": false, + "reason": "Different products and code paths: BarkProcessor voice_preset vs Qwen3 auto device mapping cudaErrorAssert." + }, + { + "left": "issue:34567", + "right": "issue:35707", + "accept": false, + "reason": "TrainerState token accounting is unrelated to progressive generation with embeddings and cache." + }, + { + "left": "issue:43421", + "right": "issue:43927", + "accept": false, + "reason": "Both involve special-token handling, but one is runtime post-processor refresh and the other is config save/load losing custom token IDs." + }, + { + "left": "issue:44368", + "right": "issue:45357", + "accept": false, + "reason": "A warning about tie_word_embeddings is unrelated to incorrect visual encoder keys in save_pretrained." + }, + { + "left": "issue:43782", + "right": "issue:45381", + "accept": false, + "reason": "Different Qwen VL bugs: from_pretrained weight_only handling vs incorrect video vision_position_ids." + }, + { + "left": "issue:44898", + "right": "issue:45072", + "accept": false, + "reason": "Different subsystems: Perceiver image resolution interpolation vs dtype mismatches in SwitchTransformers/TimmWrapperModel inference." + }, + { + "left": "issue:44451", + "right": "issue:44991", + "accept": false, + "reason": "Both are model-loading/tokenizer issues, but they involve different models and different concrete tokenizer failures." + }, + { + "left": "issue:44077", + "right": "issue:44625", + "accept": false, + "reason": "PatchTSMixer post_init policy and Qwen3.5 num_labels propagation are unrelated." + }, + { + "left": "issue:44987", + "right": "issue:45042", + "accept": false, + "reason": "Different loading failures: a specific model regression vs PIL backend image processors requiring torchvision." + }, + { + "left": "issue:45759", + "right": "issue:45859", + "accept": false, + "reason": "Different Qwen 3.x issues: config unwrapping vs CPU-offload incompatibility in rotary embedding." + }, + { + "left": "issue:43531", + "right": "issue:44589", + "accept": false, + "reason": "Sliding-window behavior in Qwen3-MoE is unrelated to missing Float8 storage support." + }, + { + "left": "issue:44062", + "right": "issue:44451", + "accept": false, + "reason": "AddedToken argument duplication is a tokenizer construction bug, not the same as a model load failure for ScandiBERT." + }, + { + "left": "issue:45704", + "right": "issue:45923", + "accept": false, + "reason": "A T5 apex RMSNorm memory leak is unrelated to a missing supports_gradient_checkpointing flag." + }, + { + "left": "issue:44918", + "right": "issue:44936", + "accept": false, + "reason": "Both are trainer-related, but one is unpacking Qwen3.5 input embeddings in TRL SFT and the other is evaluate() after train() failing." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet, and the one accepted pair (44514/44545) is supported by the identical issue title and matching crash description. The other soft-edge verdicts are conservative and avoid overstating similarity." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 14, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:30064", + "issue:36010", + "issue:39692", + "issue:40444", + "issue:42886", + "issue:43295", + "issue:43322", + "issue:43335", + "issue:43381", + "issue:43421", + "issue:43525", + "issue:43531", + "issue:43540", + "issue:43550", + "issue:43596", + "issue:43618", + "issue:43653", + "issue:43697", + "issue:43716", + "issue:43720", + "issue:43746", + "issue:43756", + "issue:43819", + "issue:43844", + "issue:43873", + "issue:43994", + "issue:44008", + "issue:44186", + "issue:44220", + "issue:44610", + "issue:44661", + "issue:44671", + "issue:44843", + "issue:45072", + "issue:45081", + "issue:45183", + "issue:45356", + "issue:45381", + "issue:45479", + "issue:45588", + "issue:45663", + "issue:45685", + "issue:45704", + "issue:45759", + "issue:45834", + "issue:45859", + "issue:45907", + "issue:45923" + ], + "soft_pairs": [ + "issue:39692|issue:44610", + "issue:43421|issue:43653", + "issue:45663|issue:45923", + "issue:45588|issue:45759", + "issue:42886|issue:44843", + "issue:43381|issue:43844", + "issue:43421|issue:44186", + "issue:43994|issue:44671", + "issue:43525|issue:43531", + "issue:45704|issue:45859", + "issue:45081|issue:45356", + "issue:39692|issue:43873", + "issue:45356|issue:45907", + "issue:43746|issue:43873", + "issue:43596|issue:44220", + "issue:40444|issue:43746", + "issue:43746|issue:45381", + "issue:43540|issue:44008", + "issue:43720|issue:43819", + "issue:44220|issue:44661", + "issue:45183|issue:45479", + "issue:43618|issue:43697", + "issue:30064|issue:36010", + "issue:43653|issue:43720", + "issue:43335|issue:43756", + "issue:43716|issue:44008", + "issue:43295|issue:45834", + "issue:43550|issue:45072", + "issue:43550|issue:43819", + "issue:43322|issue:45685" + ], + "prepared_review_unit_hash": "2dba6f3d3a5c2af21c92e9e040fafe083e3a8fb068c68e513ba2029625bd9c37", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 30, + "serialized_chars": 37183, + "estimated_input_tokens": 9296, + "estimated_eval_tokens": 18848 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is mostly a grab bag of unrelated bug reports. The only clearly duplicate-like pair is the `_patch_mistral_regex` / HF_HUB_OFFLINE tokenizer-cache regression; the Mistral-family pair involving `45081` and `45356` is too weakly connected to merge as a duplicate without stronger evidence.", + "confidence": 0.72, + "canonical_issue_reason": "issue:45081 is the best representative issue because it most directly names the failing `_patch_mistral_regex` code path and describes the concrete crash, making it the clearest root-cause-oriented report in the cluster.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45081 is the strongest global issue candidate: it is the most specific, code-path-centered report and the best anchor for the Mistral-tokenizer regression subcluster.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:39692", + "right": "issue:44610", + "accept": false, + "reason": "Different model families and failure modes: SigLIP2 doc/example mismatch vs OmDet-Turbo processor size mismatch." + }, + { + "left": "issue:43421", + "right": "issue:43653", + "accept": false, + "reason": "Both involve tokenizers, but one is runtime special-token/post-processor handling and the other is BigBird mask-token registration/empty decode output." + }, + { + "left": "issue:45663", + "right": "issue:45923", + "accept": false, + "reason": "Unrelated bugs: Gemma-4 FSDP2 shared-state KeyError vs a missing capability flag on a trust-remote-code model." + }, + { + "left": "issue:45588", + "right": "issue:45759", + "accept": false, + "reason": "Separate code paths and symptoms: flash-attention sink handling vs multimodal config unwrapping." + }, + { + "left": "issue:42886", + "right": "issue:44843", + "accept": true, + "reason": "Same underlying offline tokenizer-cache failure around `_patch_mistral_regex`: unconditional hub/model-info access breaks HF_HUB_OFFLINE mode." + }, + { + "left": "issue:43381", + "right": "issue:43844", + "accept": false, + "reason": "Different training/runtime issues: eval-mode gradient checkpointing vs ZeRO-3 gradient instability." + }, + { + "left": "issue:43421", + "right": "issue:44186", + "accept": false, + "reason": "Both are tokenizer-related, but the specific bugs are different: post-processor refresh vs LayoutLMv2 NER/padding crashes." + }, + { + "left": "issue:43994", + "right": "issue:44671", + "accept": false, + "reason": "Unrelated model-specific regressions: SigLIP2 pipeline outputs vs CamemBERT masked-LM predictions." + }, + { + "left": "issue:43525", + "right": "issue:43531", + "accept": false, + "reason": "Distinct model/config bugs: missing `pad_token_id` vs sliding-window behavior." + }, + { + "left": "issue:45704", + "right": "issue:45859", + "accept": false, + "reason": "Different implementation problems: T5 RMSNorm memory leak vs Qwen3-MoE CPU-offload incompatibility." + }, + { + "left": "issue:45081", + "right": "issue:45356", + "accept": false, + "reason": "Too weakly connected for duplicate status: `45081` is a `backend_tokenizer` AttributeError in `_patch_mistral_regex`, while `45356` is a Kimi codec/warning regression. Same helper name alone is not enough, and there are no shared targets or filenames." + }, + { + "left": "issue:39692", + "right": "issue:43873", + "accept": false, + "reason": "Doc-example/model mismatch in SigLIP2 is not the same bug as quantization/offloading behavior." + }, + { + "left": "issue:45356", + "right": "issue:45907", + "accept": false, + "reason": "Kimi tokenizer regression and a `list_repo_files` revision-forwarding bug are unrelated hub/tokenizer issues." + }, + { + "left": "issue:43746", + "right": "issue:43873", + "accept": false, + "reason": "PEFT checkpoint loading and quantization/offloading are different failure modes." + }, + { + "left": "issue:43596", + "right": "issue:44220", + "accept": false, + "reason": "Unrelated issues: DeepSpeed ZeRO-3 index error vs audio feature extraction bug." + }, + { + "left": "issue:40444", + "right": "issue:43746", + "accept": false, + "reason": "Different multimodal finetuning/loading problems with no shared code-path evidence." + }, + { + "left": "issue:43746", + "right": "issue:45381", + "accept": false, + "reason": "Local PEFT checkpoint loading and Qwen video `vision_position_ids` are separate problems." + }, + { + "left": "issue:43540", + "right": "issue:44008", + "accept": false, + "reason": "Unrelated multimodal bugs: video input handling vs audio-mask variable collision." + }, + { + "left": "issue:43720", + "right": "issue:43819", + "accept": false, + "reason": "BitNet packed-weight loading and DAC latent reconstruction are different code paths." + }, + { + "left": "issue:44220", + "right": "issue:44661", + "accept": false, + "reason": "Audio feature extraction and `TOKENIZER_MAPPING_NAMES` model registration are unrelated." + }, + { + "left": "issue:45183", + "right": "issue:45479", + "accept": false, + "reason": "Input-validation error messaging and degenerate zero-loss classification behavior are different bugs." + }, + { + "left": "issue:43618", + "right": "issue:43697", + "accept": false, + "reason": "Missing CLIP attentions assignment and RTDetrV2 output drift are distinct regressions." + }, + { + "left": "issue:30064", + "right": "issue:36010", + "accept": false, + "reason": "Completely unrelated: segmentation-map processing vs a GenerationMixin import error." + }, + { + "left": "issue:43653", + "right": "issue:43720", + "accept": false, + "reason": "BigBird special-token decoding and BitNet accelerate-loading unpacking are unrelated." + }, + { + "left": "issue:43335", + "right": "issue:43756", + "accept": false, + "reason": "SwitchTransformers sparse-layer creation and Smollm3 RoPE-layer dropping are different architecture bugs." + }, + { + "left": "issue:43716", + "right": "issue:44008", + "accept": false, + "reason": "Mistral-3 dtype mismatch and Gemma 3n variable collision are separate issues." + }, + { + "left": "issue:43295", + "right": "issue:45834", + "accept": false, + "reason": "Processor/tokenizer regression and Kosmos2.5 OCR index error are unrelated." + }, + { + "left": "issue:43550", + "right": "issue:45072", + "accept": false, + "reason": "Bamba torch.compile/SDPA failure and dtype mismatch in other models are different bugs." + }, + { + "left": "issue:43550", + "right": "issue:43819", + "accept": false, + "reason": "Bamba torch.compile/SDPA and DAC latent mismatch are unrelated." + }, + { + "left": "issue:43322", + "right": "issue:45685", + "accept": false, + "reason": "Segmentation fault on Llava Next loading and an MPS histogram not implemented error are unrelated runtime issues." + } + ] + }, + "evaluator_result": { + "accept": false, + "feedback": "The cluster summary is mostly grounded, but it is not conservative enough on the Mistral-related subcluster. In particular, `issue:45081` and `issue:45356` are only loosely connected by `fix_mistral_regex`; the packet gives different failure descriptions (a `backend_tokenizer` AttributeError vs. Kimi codec/warning regression) and no shared targets or filenames, so treating them as a clear duplicate pair is overstated. The `42886`/`44843` pair is more plausible, but the overall output still accepts a soft edge that the packet does not clearly support." + }, + "evaluator_used": true, + "retried": true, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 15, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:30064", + "issue:36331", + "issue:39692", + "issue:41720", + "issue:43421", + "issue:43531", + "issue:43540", + "issue:43577", + "issue:43582", + "issue:43606", + "issue:43643", + "issue:43644", + "issue:43653", + "issue:43749", + "issue:43782", + "issue:43818", + "issue:43819", + "issue:43825", + "issue:43827", + "issue:43994", + "issue:44062", + "issue:44186", + "issue:44442", + "issue:44448", + "issue:44479", + "issue:44485", + "issue:44488", + "issue:44509", + "issue:44561", + "issue:44671", + "issue:44805", + "issue:44811", + "issue:44821", + "issue:44843", + "issue:44877", + "issue:44898", + "issue:44936", + "issue:44977", + "issue:44991", + "issue:45020", + "issue:45072", + "issue:45081", + "issue:45084", + "issue:45381", + "issue:45584", + "issue:45701", + "issue:45834" + ], + "soft_pairs": [ + "issue:44062|issue:44488", + "issue:44186|issue:44898", + "issue:44843|issue:45081", + "issue:43531|issue:43643", + "issue:43653|issue:43819", + "issue:43577|issue:43818", + "issue:43643|issue:44877", + "issue:43540|issue:45381", + "issue:43421|issue:44811", + "issue:43782|issue:44977", + "issue:44448|issue:44671", + "issue:45584|issue:45834", + "issue:43825|issue:43994", + "issue:44561|issue:45020", + "issue:39692|issue:45834", + "issue:44485|issue:44509", + "issue:44821|issue:45084", + "issue:30064|issue:36331", + "issue:41720|issue:44991", + "issue:43644|issue:43749", + "issue:44186|issue:44442", + "issue:43825|issue:43827", + "issue:44479|issue:45834", + "issue:43577|issue:43582", + "issue:44805|issue:44936", + "issue:43606|issue:45072", + "issue:44671|issue:45701" + ], + "prepared_review_unit_hash": "302a991f7d2001395c60bed868515b7b351c9731cb7178af2811b38e368edab4", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 27, + "serialized_chars": 35265, + "estimated_input_tokens": 8817, + "estimated_eval_tokens": 17890 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is mostly a set of unrelated bug reports that only share broad themes like tokenizer, multimodal, and v5 regression language. The soft pairs are mostly superficial similarities and do not look like the same underlying bug or change.", + "confidence": 0.9, + "canonical_issue_reason": "issue:45020 is the broadest open umbrella here, covering recent version regressions around remote_code/config loading; the other issues are narrower, model-specific failures.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45020 is the best anchor for the cluster because it is the most general and central regression report, even though it does not duplicate the other issues.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44062", + "right": "issue:44488", + "accept": false, + "reason": "Different failure modes: AddedToken constructor error vs tokenizer/model loading failure for a specific repo." + }, + { + "left": "issue:44186", + "right": "issue:44898", + "accept": false, + "reason": "Different models and code paths: LayoutLMv2 tokenizer padding/NER crash vs Perceiver image classification resolution/interpolation bug." + }, + { + "left": "issue:44843", + "right": "issue:45081", + "accept": false, + "reason": "Both touch _patch_mistral_regex, but one is an offline model_info call and the other is an AttributeError on tokenizer internals; not the same bug." + }, + { + "left": "issue:43531", + "right": "issue:43643", + "accept": false, + "reason": "Unrelated areas: Qwen3-MoE sliding_window behavior vs AutoConfig trust_remote_code field loss." + }, + { + "left": "issue:43653", + "right": "issue:43819", + "accept": false, + "reason": "Tokenizer special-token registration bug vs DAC latents/STE forward mismatch; no common code-path." + }, + { + "left": "issue:43577", + "right": "issue:43818", + "accept": false, + "reason": "Different multimodal models and defects: dtype propagation in Blip2 vs missing temporal attention / shared weights in Video-LLaVA." + }, + { + "left": "issue:43643", + "right": "issue:44877", + "accept": false, + "reason": "Both involve config loading, but one is missing fields from trust_remote_code and the other is strict config rejection for granite_speech." + }, + { + "left": "issue:43540", + "right": "issue:45381", + "accept": false, + "reason": "Different Qwen video bugs: Qwen3OmniMoe video processing ValueError vs Qwen2.5-VL vision_position_ids regression." + }, + { + "left": "issue:43421", + "right": "issue:44811", + "accept": false, + "reason": "Runtime special-token/post-processor mutation issue vs Whisper batch_decode skip_special_tokens handling; different components." + }, + { + "left": "issue:43782", + "right": "issue:44977", + "accept": false, + "reason": "Loading error with weight_only=True vs generation failure with flash-attention; not the same concrete bug." + }, + { + "left": "issue:44448", + "right": "issue:44671", + "accept": false, + "reason": "Both are v5 regressions, but one is Pegasus output drift and the other is CamemBERT masked-LM prediction corruption; different models and causes." + }, + { + "left": "issue:45584", + "right": "issue:45834", + "accept": false, + "reason": "Whisper empty transcription generation vs Kosmos2.5 long OCR index error; different models and failure signatures." + }, + { + "left": "issue:43825", + "right": "issue:43994", + "accept": false, + "reason": "Docs/error-message wording around removed pipeline tasks vs a SigLIP2 model/pipeline output problem; not the same bug." + }, + { + "left": "issue:44561", + "right": "issue:45020", + "accept": false, + "reason": "Related only at a broad level: 44561 is a specific remote_code break from removing is_torch_fx_available, while 45020 is a generic umbrella report." + }, + { + "left": "issue:39692", + "right": "issue:45834", + "accept": false, + "reason": "SigLIP2 docs/example errors are unrelated to Kosmos2.5 OCR index failures." + }, + { + "left": "issue:44485", + "right": "issue:44509", + "accept": false, + "reason": "GLM-5 RoPE implementation discussion vs outdated docs about removed pipeline tasks; different issues." + }, + { + "left": "issue:44821", + "right": "issue:45084", + "accept": false, + "reason": "AutoImageProcessor URL loading vs Jinja/template compilation TypeError; unrelated subsystems." + }, + { + "left": "issue:30064", + "right": "issue:36331", + "accept": false, + "reason": "Image processor segmentation-map failure vs trainer.compute_loss signature mismatch; no overlap in code path." + }, + { + "left": "issue:41720", + "right": "issue:44991", + "accept": false, + "reason": "Qwen3 auto device-map CUDA assert vs tokenizer loading regression for EMBEDDIA/est-roberta; different problems." + }, + { + "left": "issue:43644", + "right": "issue:43749", + "accept": false, + "reason": "Non-persistent buffer initialization bug vs FSDP CPU RAM efficient loading breakage; related only by version." + }, + { + "left": "issue:44186", + "right": "issue:44442", + "accept": false, + "reason": "Tokenizer crash in LayoutLMv2 vs AutoTokenizer loading failure for FastSpeech2Conformer; different artifacts and code paths." + }, + { + "left": "issue:43825", + "right": "issue:43827", + "accept": false, + "reason": "Both reference removed v5 pipeline tasks, but one is a runtime message issue and the other is stale docs; not the same bug/change." + }, + { + "left": "issue:44479", + "right": "issue:45834", + "accept": false, + "reason": "Qwen video-input regression vs Kosmos2.5 OCR long-input index error; unrelated model families." + }, + { + "left": "issue:43577", + "right": "issue:43582", + "accept": false, + "reason": "Blip2 dtype propagation issue vs Apple Silicon caching_allocator_warmup TypeError; distinct failures." + }, + { + "left": "issue:44805", + "right": "issue:44936", + "accept": false, + "reason": "Tensor mask shape mismatch vs trainer.evaluate after train lifecycle bug; different concrete problems." + }, + { + "left": "issue:43606", + "right": "issue:45072", + "accept": false, + "reason": "CPU offload device mismatch in bark-small vs bfloat16 dtype mismatches in SwitchTransformers/TimmWrapperModel." + }, + { + "left": "issue:44671", + "right": "issue:45701", + "accept": false, + "reason": "CamemBERT v5 prediction regression vs general tokenization changes across versions; not the same bug." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 16, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:34634", + "issue:36010", + "issue:36331", + "issue:38175", + "issue:39692", + "issue:42175", + "issue:43257", + "issue:43262", + "issue:43329", + "issue:43452", + "issue:43526", + "issue:43582", + "issue:43638", + "issue:43650", + "issue:43653", + "issue:43697", + "issue:43720", + "issue:43742", + "issue:43746", + "issue:43749", + "issue:43782", + "issue:43819", + "issue:43854", + "issue:43906", + "issue:43927", + "issue:44060", + "issue:44291", + "issue:44368", + "issue:44554", + "issue:44589", + "issue:44610", + "issue:44811", + "issue:44857", + "issue:44912", + "issue:44936", + "issue:44977", + "issue:44991", + "issue:45020", + "issue:45042", + "issue:45072", + "issue:45081", + "issue:45084", + "issue:45325", + "issue:45406", + "issue:45459", + "issue:45538", + "issue:45600", + "issue:45803" + ], + "soft_pairs": [ + "issue:34634|issue:36010", + "issue:44610|issue:45072", + "issue:43782|issue:45325", + "issue:43720|issue:45072", + "issue:42175|issue:45081", + "issue:45081|issue:45406", + "issue:43749|issue:44291", + "issue:43650|issue:43906", + "issue:44060|issue:44368", + "issue:43582|issue:44857", + "issue:43262|issue:43819", + "issue:43526|issue:43697", + "issue:43526|issue:44554", + "issue:45459|issue:45803", + "issue:34634|issue:36331", + "issue:45084|issue:45538", + "issue:43329|issue:45600", + "issue:43742|issue:45020", + "issue:43653|issue:44912", + "issue:43638|issue:44936", + "issue:43746|issue:44811", + "issue:43452|issue:44991", + "issue:38175|issue:44610", + "issue:44589|issue:45042", + "issue:43854|issue:44977", + "issue:43819|issue:43927", + "issue:39692|issue:44912", + "issue:43257|issue:43329" + ], + "prepared_review_unit_hash": "3cfa70e286fe3ffcd2c7c02b2e87b8a7628b57cd9fe1d17b595c5a62bf7a38fb", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 28, + "serialized_chars": 36472, + "estimated_input_tokens": 9118, + "estimated_eval_tokens": 18492 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Mostly unrelated Transformers bug reports spanning model loading, tokenizers, processors, quantization, and training. None of the soft pairs look like the same underlying defect, so all candidate merges are rejected. The broadest representative issue is the remote_code regression report.", + "confidence": 0.86, + "canonical_issue_reason": "issue:45020 is the broadest umbrella report about recent Transformers versions breaking remote_code models, so it best represents this mixed regression cluster.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45020 is the most general and centrally scoped issue in the set; the others are narrower, model-specific failures rather than a shared duplicate.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:34634", + "right": "issue:36010", + "accept": false, + "reason": "Different subsystems and failure modes: BarkProcessor voice_preset vs GenerationMixin import regression." + }, + { + "left": "issue:44610", + "right": "issue:45072", + "accept": false, + "reason": "Different bug classes: processor image size mismatch vs dtype mismatch in inference." + }, + { + "left": "issue:43782", + "right": "issue:45325", + "accept": false, + "reason": "Both are model-loading related, but one is weight_only loading and the other is a rope-index temporal-position bug." + }, + { + "left": "issue:43720", + "right": "issue:45072", + "accept": false, + "reason": "BitNet packed-weight loading failure is unrelated to bfloat16 dtype mismatches in SwitchTransformers/TimmWrapperModel." + }, + { + "left": "issue:42175", + "right": "issue:45081", + "accept": false, + "reason": "Packaging/backend dependency issue vs tokenizer regex patch crash; not the same defect." + }, + { + "left": "issue:45081", + "right": "issue:45406", + "accept": false, + "reason": "Both involve tokenizer/processor code, but different objects and crashes in different call paths." + }, + { + "left": "issue:43749", + "right": "issue:44291", + "accept": false, + "reason": "Both concern loading flows, but one is FSDP CPU RAM-efficient loading and the other is init_empty_weights argument handling." + }, + { + "left": "issue:43650", + "right": "issue:43906", + "accept": false, + "reason": "Generic placeholder issue vs isolated reproduction of a specific external bug; not duplicates." + }, + { + "left": "issue:44060", + "right": "issue:44368", + "accept": false, + "reason": "Different tied-weights warnings affecting different model families and configurations." + }, + { + "left": "issue:43582", + "right": "issue:44857", + "accept": false, + "reason": "Apple Silicon allocator TypeError is unrelated to CUDA AMP crash in a detection loss." + }, + { + "left": "issue:43262", + "right": "issue:43819", + "accept": false, + "reason": "Audio sampling-rate default bug is unrelated to DAC latent/forward-pass mismatch." + }, + { + "left": "issue:43526", + "right": "issue:43697", + "accept": false, + "reason": "BEiT reduce_labels bug and RTDetrV2 output drift are separate model-specific defects." + }, + { + "left": "issue:43526", + "right": "issue:44554", + "accept": false, + "reason": "No shared code path: BEiT label reduction vs MPS attention correctness on differing head dims." + }, + { + "left": "issue:45459", + "right": "issue:45803", + "accept": false, + "reason": "Both involve error handling, but one is protobuf decode error masking and the other is a bare except in tensor conversion." + }, + { + "left": "issue:34634", + "right": "issue:36331", + "accept": false, + "reason": "Unrelated: BarkProcessor preset behavior vs CustomTrainer compute_loss signature change." + }, + { + "left": "issue:45084", + "right": "issue:45538", + "accept": false, + "reason": "Template compilation error and CLIPTokenizer max length bug are unrelated." + }, + { + "left": "issue:43329", + "right": "issue:45600", + "accept": false, + "reason": "Multimodal token counting bug is unrelated to stale auto_mappings config references." + }, + { + "left": "issue:43742", + "right": "issue:45020", + "accept": false, + "reason": "Specific MobileLLM key error does not match the broader remote_code regression." + }, + { + "left": "issue:43653", + "right": "issue:44912", + "accept": false, + "reason": "Tokenizer special-token registration bug is unrelated to MXFP4 quantization fallback." + }, + { + "left": "issue:43638", + "right": "issue:44936", + "accept": false, + "reason": "Zero3/non-pretrained Bert index error is not the same as trainer.evaluate() failing after train()." + }, + { + "left": "issue:43746", + "right": "issue:44811", + "accept": false, + "reason": "Local PEFT checkpoint loading issue is unrelated to Whisper batch_decode skip_special_tokens handling." + }, + { + "left": "issue:43452", + "right": "issue:44991", + "accept": false, + "reason": "gguf_file support failure is different from tokenizer loading regression for est-roberta." + }, + { + "left": "issue:38175", + "right": "issue:44610", + "accept": false, + "reason": "Unexpected zero probabilities in SigLIP2 are not the same as OmDet processor input-size mismatch." + }, + { + "left": "issue:44589", + "right": "issue:45042", + "accept": false, + "reason": "Storage lookup failure for float8 is unrelated to torchvision dependency requirements in PIL image processors." + }, + { + "left": "issue:43854", + "right": "issue:44977", + "accept": false, + "reason": "Model loading/test issue for GLM-4.7-Flash is not the same as flash-attention generation failure in Qwen3.5." + }, + { + "left": "issue:43819", + "right": "issue:43927", + "accept": false, + "reason": "DAC latent mismatch and DiaConfig custom token ID loss are separate generation/model-config bugs." + }, + { + "left": "issue:39692", + "right": "issue:44912", + "accept": false, + "reason": "SigLIP2 doc example errors do not match MXFP4 quantization fallback behavior." + }, + { + "left": "issue:43257", + "right": "issue:43329", + "accept": false, + "reason": "Qwen3 MOE weight conversion/loading issue is unrelated to multimodal token counting undefined variables." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 17, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:33357", + "issue:34567", + "issue:36010", + "issue:36331", + "issue:40990", + "issue:41720", + "issue:42490", + "issue:42915", + "issue:43257", + "issue:43262", + "issue:43329", + "issue:43450", + "issue:43452", + "issue:43454", + "issue:43531", + "issue:43540", + "issue:43550", + "issue:43577", + "issue:43582", + "issue:43606", + "issue:43643", + "issue:43645", + "issue:43696", + "issue:43697", + "issue:43720", + "issue:43742", + "issue:43749", + "issue:43756", + "issue:43931", + "issue:44291", + "issue:44479", + "issue:44568", + "issue:44610", + "issue:44756", + "issue:44811", + "issue:44849", + "issue:44912", + "issue:44987", + "issue:45042", + "issue:45356", + "issue:45357", + "issue:45381", + "issue:45488", + "issue:45538", + "issue:45701", + "issue:45759", + "issue:45859", + "issue:45907" + ], + "soft_pairs": [ + "issue:36010|issue:43645", + "issue:43720|issue:44610", + "issue:43550|issue:44912", + "issue:43454|issue:43577", + "issue:43257|issue:43696", + "issue:43262|issue:44811", + "issue:43329|issue:44479", + "issue:42490|issue:43257", + "issue:43531|issue:45042", + "issue:36010|issue:36331", + "issue:44568|issue:45356", + "issue:43749|issue:44756", + "issue:43643|issue:45759", + "issue:44568|issue:45488", + "issue:34567|issue:40990", + "issue:43329|issue:43450", + "issue:44849|issue:45701", + "issue:43257|issue:43931", + "issue:43606|issue:45859", + "issue:43697|issue:45357", + "issue:43257|issue:43756", + "issue:44479|issue:45381", + "issue:43452|issue:43582", + "issue:44987|issue:45538", + "issue:41720|issue:43540", + "issue:44291|issue:45042", + "issue:42915|issue:43742", + "issue:44987|issue:45907", + "issue:33357|issue:34567" + ], + "prepared_review_unit_hash": "3282751fd1ac6e24f64b39d934f4103215310d01de41d429960799e10ff376e0", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 29, + "serialized_chars": 36597, + "estimated_input_tokens": 9150, + "estimated_eval_tokens": 18556 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is heterogeneous: tokenization regressions, model-loading bugs, multimodal/video processor issues, and training API breakages. None of the soft pairs look like the same underlying bug/change, so I would not deduplicate them.", + "confidence": 0.94, + "canonical_issue_reason": "issue:45701 is the broadest open issue and best fits an umbrella for the tokenizer/version-regression subset, with multiple inbound references and a title that captures the general symptom pattern.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45701 is the most central and reusable issue in this set; it\u2019s broad enough to represent the shared tokenizer-regression theme better than the highly specific model- or API-specific reports.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:36010", + "right": "issue:43645", + "accept": false, + "reason": "Different breakages: a missing GenerationMixin import vs Jupyter-defined custom model initialization in v5." + }, + { + "left": "issue:43720", + "right": "issue:44610", + "accept": false, + "reason": "Both are loading/vision-adjacent failures, but one is about unpacking packed BitNet weights during accelerate loading and the other about processor image size mismatch." + }, + { + "left": "issue:43550", + "right": "issue:44912", + "accept": false, + "reason": "Different concrete problems: torch.compile with SDPA on Bamba vs MXFP4 quantization fallback on gpt-oss-20b." + }, + { + "left": "issue:43454", + "right": "issue:43577", + "accept": false, + "reason": "AyaVision weight tying bug is unrelated to Blip2 dtype propagation on model load." + }, + { + "left": "issue:43257", + "right": "issue:43696", + "accept": false, + "reason": "Qwen3 MOE weights not converted during accelerate+deepspeed is distinct from a plain CUDA OOM on gpt-oss-20b." + }, + { + "left": "issue:43262", + "right": "issue:44811", + "accept": false, + "reason": "Audio apply_chat_template sampling-rate default bug is unrelated to Whisper batch_decode skipping special tokens." + }, + { + "left": "issue:43329", + "right": "issue:44479", + "accept": false, + "reason": "Undefined video token counting helpers vs a broader video-input regression in several VLMs are related by modality only, not the same bug." + }, + { + "left": "issue:42490", + "right": "issue:43257", + "accept": false, + "reason": "Local-path import warning with trailing slash is unrelated to Qwen3 MOE weight conversion under accelerate/deepspeed." + }, + { + "left": "issue:43531", + "right": "issue:45042", + "accept": false, + "reason": "Sliding window behavior in Qwen3-MoE is not the same as PIL-backed image processors wrongly requiring torchvision." + }, + { + "left": "issue:36010", + "right": "issue:36331", + "accept": false, + "reason": "ImportError for GenerationMixin and unexpected num_items_in_batch in CustomTrainer are separate API regressions." + }, + { + "left": "issue:44568", + "right": "issue:45356", + "accept": false, + "reason": "Both tokenizer-related, but one is add_special_tokens/BOS-EOS behavior while the other is codec handling and warning text in Kimi-K2.5." + }, + { + "left": "issue:43749", + "right": "issue:44756", + "accept": false, + "reason": "FSDP CPU RAM efficient loading and disabling mmap on Strix Halo are different memory-loading issues with different triggers." + }, + { + "left": "issue:43643", + "right": "issue:45759", + "accept": false, + "reason": "Missing fields from trust_remote_code AutoConfig is not the same as AutoModelForCausalLM.from_config failing to unwrap text_config in composite Qwen configs." + }, + { + "left": "issue:44568", + "right": "issue:45488", + "accept": false, + "reason": "Different tokenizer regressions: add_special_tokens handling vs LlamaTokenizer overriding ByteLevel pre-tokenizer." + }, + { + "left": "issue:34567", + "right": "issue:40990", + "accept": false, + "reason": "TrainerState token counting not updating is unrelated to high perplexity on a specific model/dataset." + }, + { + "left": "issue:43329", + "right": "issue:43450", + "accept": false, + "reason": "Undefined helper in multimodal token counting vs batched video processor shape mismatch are separate video-path bugs." + }, + { + "left": "issue:44849", + "right": "issue:45701", + "accept": false, + "reason": "Qwen3.5 output_hidden_states bug is model-specific and not the same as the broad tokenization-version regression." + }, + { + "left": "issue:43257", + "right": "issue:43931", + "accept": false, + "reason": "Qwen3 MOE accelerate/deepspeed conversion failure is different from Qwen3-VL weight shape mismatch on model load." + }, + { + "left": "issue:43606", + "right": "issue:45859", + "accept": false, + "reason": "CPU offload device mismatch in bark-small is unrelated to Qwen3_5MoeTextRotaryEmbedding CPU offload compatibility." + }, + { + "left": "issue:43697", + "right": "issue:45357", + "accept": false, + "reason": "RTDetrV2 differing outputs in v5 is not the same as Qwen3.5 save_pretrained writing incorrect visual encoder keys." + }, + { + "left": "issue:43257", + "right": "issue:43756", + "accept": false, + "reason": "Qwen3 MOE conversion/loading failure is unrelated to Smollm3 RoPE layer dropping." + }, + { + "left": "issue:44479", + "right": "issue:45381", + "accept": false, + "reason": "General video regression across several models is too broad to merge with the specific qwen2.5-vl vision_position_ids issue." + }, + { + "left": "issue:43452", + "right": "issue:43582", + "accept": false, + "reason": "gguf_file breaks AutoTokenizer/AutoModel loading, while caching_allocator_warmup fails on Apple Silicon with TypeError; different code paths." + }, + { + "left": "issue:44987", + "right": "issue:45538", + "accept": false, + "reason": "Physical-intelligence/fast loading failure is unrelated to CLIPTokenizer's pathological model_max_length." + }, + { + "left": "issue:41720", + "right": "issue:43540", + "accept": false, + "reason": "Qwen3 auto device mapping cudaErrorAssert and Qwen3OmniMoe video input ValueError are different runtime failures." + }, + { + "left": "issue:44291", + "right": "issue:45042", + "accept": false, + "reason": "init_empty_weights _is_hf_initialized argument issue is unrelated to PIL backend processors requiring torchvision." + }, + { + "left": "issue:42915", + "right": "issue:43742", + "accept": false, + "reason": "Qwen3Moe FineGrainedFP8Config failure and MobileLLM-125M key error are distinct loading/config bugs." + }, + { + "left": "issue:44987", + "right": "issue:45907", + "accept": false, + "reason": "Loading physical-intelligence/fast on transformers>=5.1.0 is unrelated to list_repo_files not forwarding revision in CI." + }, + { + "left": "issue:33357", + "right": "issue:34567", + "accept": false, + "reason": "MacOS bus error on CLIP model load is unrelated to TrainerState token-count tracking." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 18, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:28282", + "issue:29127", + "issue:34567", + "issue:34689", + "issue:36331", + "issue:38175", + "issue:42175", + "issue:42915", + "issue:43317", + "issue:43452", + "issue:43493", + "issue:43526", + "issue:43540", + "issue:43577", + "issue:43582", + "issue:43643", + "issue:43742", + "issue:43749", + "issue:43756", + "issue:43782", + "issue:43824", + "issue:43906", + "issue:43927", + "issue:43994", + "issue:44062", + "issue:44291", + "issue:44479", + "issue:44488", + "issue:44560", + "issue:44737", + "issue:44821", + "issue:44898", + "issue:44912", + "issue:44938", + "issue:44991", + "issue:45072", + "issue:45081", + "issue:45137", + "issue:45325", + "issue:45357", + "issue:45375", + "issue:45381", + "issue:45405", + "issue:45479", + "issue:45538", + "issue:45588", + "issue:45823", + "issue:45920" + ], + "soft_pairs": [ + "issue:42915|issue:44821", + "issue:42915|issue:45538", + "issue:43927|issue:44912", + "issue:43317|issue:45538", + "issue:42175|issue:43824", + "issue:43540|issue:44479", + "issue:43540|issue:44560", + "issue:43577|issue:45072", + "issue:44062|issue:44991", + "issue:42175|issue:44938", + "issue:38175|issue:44898", + "issue:42175|issue:44488", + "issue:45137|issue:45405", + "issue:43906|issue:45381", + "issue:44737|issue:45588", + "issue:28282|issue:34689", + "issue:42915|issue:43749", + "issue:43742|issue:44291", + "issue:43526|issue:45479", + "issue:43582|issue:43782", + "issue:43540|issue:45823", + "issue:43756|issue:45325", + "issue:28282|issue:29127", + "issue:43493|issue:43994", + "issue:34567|issue:36331", + "issue:44062|issue:45081", + "issue:44291|issue:44821", + "issue:45357|issue:45920", + "issue:43643|issue:45375", + "issue:43452|issue:44488" + ], + "prepared_review_unit_hash": "27922551fc9691a35b2ee5705e8d84f5adb6ab6923711f46e94d071bcf194b92", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 30, + "serialized_chars": 37039, + "estimated_input_tokens": 9260, + "estimated_eval_tokens": 18776 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Most items are unrelated bug reports across different models/subsystems. I found one plausible duplicate pair around SigLIP2; the rest should stay separate.", + "confidence": 0.87, + "canonical_issue_reason": "No single canonical issue for the full set; the only believable duplicate pair is the SigLIP2 reports, where 43994 is the clearer anchor because it captures the user-visible failure and broader impact.", + "canonical_pr_reason": null, + "best_issue_reason": "43994 is the best overall representative among these issues because it describes a broad, user-facing regression in a widely used model path, while most others are narrower or unrelated.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:42915", + "right": "issue:44821", + "accept": false, + "reason": "Different models and failures: Qwen3Moe FP8 loading vs AutoImageProcessor URL loading." + }, + { + "left": "issue:42915", + "right": "issue:45538", + "accept": false, + "reason": "Unrelated bugs: Qwen3Moe FP8 config error vs CLIPTokenizer max_length behavior." + }, + { + "left": "issue:43927", + "right": "issue:44912", + "accept": false, + "reason": "Different code paths: config save/load token IDs vs MXFP4 quantization fallback." + }, + { + "left": "issue:43317", + "right": "issue:45538", + "accept": false, + "reason": "Device-map/offload loading bug is unrelated to CLIPTokenizer max_length." + }, + { + "left": "issue:42175", + "right": "issue:43824", + "accept": false, + "reason": "Backend packaging issue vs missing Qwen2.5-VL class import." + }, + { + "left": "issue:43540", + "right": "issue:44479", + "accept": false, + "reason": "Both are video-related, but they target different Qwen model families and failure modes; not enough evidence of one underlying bug." + }, + { + "left": "issue:43540", + "right": "issue:44560", + "accept": false, + "reason": "Different video-input errors in different models; not the same concrete bug." + }, + { + "left": "issue:43577", + "right": "issue:45072", + "accept": false, + "reason": "Both mention dtype, but they affect different architectures and symptoms." + }, + { + "left": "issue:44062", + "right": "issue:44991", + "accept": false, + "reason": "Tokenizer errors, but one is AddedToken argument handling and the other is loading a specific tokenizer." + }, + { + "left": "issue:42175", + "right": "issue:44938", + "accept": false, + "reason": "Backend installation mismatch is unrelated to Python 3.14 import/load failure." + }, + { + "left": "issue:38175", + "right": "issue:44898", + "accept": false, + "reason": "Different models and failures: SigLIP2 zero probabilities vs Perceiver non-default resolution." + }, + { + "left": "issue:42175", + "right": "issue:44488", + "accept": false, + "reason": "Packaging/backend issue vs model loading failure for a specific checkpoint." + }, + { + "left": "issue:45137", + "right": "issue:45405", + "accept": false, + "reason": "DeepSpeed ZeRO3 runtime error is unrelated to PEFT version bump." + }, + { + "left": "issue:43906", + "right": "issue:45381", + "accept": false, + "reason": "No clear shared bug: isolated reproduction of an old issue vs Qwen2.5-VL video position IDs." + }, + { + "left": "issue:44737", + "right": "issue:45588", + "accept": false, + "reason": "XLNet CPU placement bug is unrelated to flash_attention s_aux=None crash." + }, + { + "left": "issue:28282", + "right": "issue:34689", + "accept": false, + "reason": "Generic PyTorch ImportError vs a specific Llama 3.2 Vision loading regression." + }, + { + "left": "issue:42915", + "right": "issue:43749", + "accept": false, + "reason": "Different training/loading problems with no shared failing code path." + }, + { + "left": "issue:43742", + "right": "issue:44291", + "accept": false, + "reason": "Both are load-time issues, but one is a missing key in MobileLLM and the other is an unexpected init_empty_weights argument." + }, + { + "left": "issue:43526", + "right": "issue:45479", + "accept": false, + "reason": "Image processor label reduction bug vs sequence-classification zero-loss bug." + }, + { + "left": "issue:43582", + "right": "issue:43782", + "accept": false, + "reason": "Apple Silicon allocator TypeError is unrelated to Qwen3VL weight_only loading failure." + }, + { + "left": "issue:43540", + "right": "issue:45823", + "accept": false, + "reason": "Video processing error vs Gemma4 PLE device mismatch; different subsystems." + }, + { + "left": "issue:43756", + "right": "issue:45325", + "accept": false, + "reason": "Different RoPE-related model bugs with different symptoms and affected models." + }, + { + "left": "issue:28282", + "right": "issue:29127", + "accept": false, + "reason": "PyTorch missing import error is unrelated to LayoutLMv3 box-validation messaging." + }, + { + "left": "issue:43493", + "right": "issue:43994", + "accept": true, + "reason": "Both are SigLIP2 regressions describing the same underlying model discrepancy, with 43994 showing the user-facing symptom and 43493 giving the implementation-vs-JAX diagnosis." + }, + { + "left": "issue:34567", + "right": "issue:36331", + "accept": false, + "reason": "TrainerState token counting bug vs CustomTrainer compute_loss signature mismatch." + }, + { + "left": "issue:44062", + "right": "issue:45081", + "accept": false, + "reason": "Different tokenizer failures: AddedToken special-arg duplication vs Mistral regex patch crash." + }, + { + "left": "issue:44291", + "right": "issue:44821", + "accept": false, + "reason": "Model init-empty-weights argument bug is unrelated to AutoImageProcessor URL loading." + }, + { + "left": "issue:45357", + "right": "issue:45920", + "accept": false, + "reason": "save_pretrained visual-key regression is not the same as AutoTokenizer wrong token IDs." + }, + { + "left": "issue:43643", + "right": "issue:45375", + "accept": false, + "reason": "Both involve config fields, but one is trust_remote_code AutoConfig loss and the other is a specific Qwen vision config field." + }, + { + "left": "issue:43452", + "right": "issue:44488", + "accept": false, + "reason": "GGUF tokenizer/model loading breakage is unrelated to the sleng-bert loading report." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet and stays conservative. It correctly treats most pairs as unrelated and only tentatively groups the SigLIP2 issues, which is the one pair whose titles strongly align. I don\u2019t see any overstatement that would require rejection." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 19, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:24643", + "issue:29127", + "issue:30064", + "issue:33666", + "issue:38175", + "issue:41762", + "issue:43295", + "issue:43452", + "issue:43531", + "issue:43577", + "issue:43643", + "issue:43653", + "issue:43688", + "issue:43697", + "issue:43720", + "issue:43742", + "issue:43749", + "issue:43819", + "issue:44361", + "issue:44442", + "issue:44464", + "issue:44466", + "issue:44484", + "issue:44514", + "issue:44561", + "issue:44568", + "issue:44589", + "issue:44625", + "issue:44704", + "issue:44743", + "issue:44792", + "issue:44811", + "issue:44821", + "issue:44843", + "issue:44877", + "issue:44898", + "issue:44936", + "issue:44964", + "issue:45005", + "issue:45042", + "issue:45081", + "issue:45216", + "issue:45290", + "issue:45356", + "issue:45381", + "issue:45405", + "issue:45600", + "issue:45823" + ], + "soft_pairs": [ + "issue:44792|issue:44964", + "issue:45042|issue:45081", + "issue:43742|issue:44589", + "issue:43531|issue:44821", + "issue:45405|issue:45600", + "issue:43697|issue:45216", + "issue:43688|issue:43749", + "issue:44514|issue:45381", + "issue:43720|issue:44898", + "issue:44484|issue:44704", + "issue:41762|issue:44936", + "issue:24643|issue:29127", + "issue:44625|issue:44877", + "issue:44514|issue:45290", + "issue:43452|issue:43577", + "issue:43452|issue:44843", + "issue:44466|issue:45005", + "issue:43819|issue:44811", + "issue:43295|issue:45356", + "issue:44936|issue:45823", + "issue:43653|issue:44568", + "issue:44464|issue:44743", + "issue:38175|issue:43295", + "issue:43643|issue:44561", + "issue:43653|issue:44361", + "issue:44361|issue:44442", + "issue:30064|issue:33666" + ], + "prepared_review_unit_hash": "ed851354ef47701a33e732a66e00c4f0b7509399f5561e92cac6613d8ff85ae2", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 27, + "serialized_chars": 35983, + "estimated_input_tokens": 8996, + "estimated_eval_tokens": 18248 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Most items are unrelated single-issue reports. The only plausible duplicate/near-duplicate pair is the tied-weights save/serialization regression in v5. No PRs are present.", + "confidence": 0.74, + "canonical_issue_reason": "issue:44466 is the broadest, most concrete report in the only likely duplicate family: it describes the tied-weights serialization problem in v5 in a way that can absorb model-specific variants.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:44466 is the best representative issue for triage because it focuses on the underlying save_pretrained/tied-weights regression rather than a narrow model instance.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44792", + "right": "issue:44964", + "accept": false, + "reason": "Different failures: a janus image-generation test case versus loading Phi-4 multimodal models. No shared code path." + }, + { + "left": "issue:45042", + "right": "issue:45081", + "accept": false, + "reason": "Unrelated subsystems: PIL image processor torchvision dependency vs Mistral regex/tokenizer loading crash." + }, + { + "left": "issue:43742", + "right": "issue:44589", + "accept": false, + "reason": "Different model/load-time errors: MobileLLM key error versus Float8 storage lookup failure." + }, + { + "left": "issue:43531", + "right": "issue:44821", + "accept": false, + "reason": "Qwen3-MoE sliding_window behavior is unrelated to loading AutoImageProcessor from a URL." + }, + { + "left": "issue:45405", + "right": "issue:45600", + "accept": false, + "reason": "One is a dependency version bump issue; the other is stale auto_mappings references to removed configs." + }, + { + "left": "issue:43697", + "right": "issue:45216", + "accept": false, + "reason": "Different bugs: RTDetrV2 output changes versus Qwen3.5 save_pretrained regression." + }, + { + "left": "issue:43688", + "right": "issue:43749", + "accept": false, + "reason": "Both involve training/runtime infrastructure, but one is auxiliary-loss normalization and the other is FSDP loading." + }, + { + "left": "issue:44514", + "right": "issue:45381", + "accept": false, + "reason": "Different chat-template/video-position-id problems; not the same concrete bug." + }, + { + "left": "issue:43720", + "right": "issue:44898", + "accept": false, + "reason": "BitNet accelerated-loading unpacking bug versus Perceiver image classification interpolation failure." + }, + { + "left": "issue:44484", + "right": "issue:44704", + "accept": false, + "reason": "save_pretrained shard size question is unrelated to AutoProcessor kwargs forwarding." + }, + { + "left": "issue:41762", + "right": "issue:44936", + "accept": false, + "reason": "Gemma3 DeepSpeed ZeRO-3 load failure is not the same as trainer.evaluate() failing after train()." + }, + { + "left": "issue:24643", + "right": "issue:29127", + "accept": false, + "reason": "DeepSpeed weight-dimension runtime error is unrelated to LayoutLMv3 box validation messaging." + }, + { + "left": "issue:44625", + "right": "issue:44877", + "accept": false, + "reason": "Qwen3.5 num_labels propagation is unrelated to granite_speech strict-config loading." + }, + { + "left": "issue:44514", + "right": "issue:45290", + "accept": false, + "reason": "Both are apply_chat_template-related, but they fail on different inputs and code paths." + }, + { + "left": "issue:43452", + "right": "issue:43577", + "accept": false, + "reason": "Tokenizer/model loading via gguf_file is unrelated to Blip2 dtype propagation." + }, + { + "left": "issue:43452", + "right": "issue:44843", + "accept": false, + "reason": "Different tokenizer-loading failures: gguf_file support versus offline model_info lookup in mistral regex patching." + }, + { + "left": "issue:44466", + "right": "issue:45005", + "accept": true, + "reason": "Both report the same tied-weights regression in v5 save/serialization behavior, with 45005 appearing model-specific and 44466 describing the underlying device-dependent bug." + }, + { + "left": "issue:43819", + "right": "issue:44811", + "accept": false, + "reason": "DAC latent conversion mismatch and Whisper batch_decode special-token handling are unrelated." + }, + { + "left": "issue:43295", + "right": "issue:45356", + "accept": false, + "reason": "Both mention tokenizer/processor regressions, but one is custom processor.tokenizer access while the other is Kimi codec handling and warning behavior." + }, + { + "left": "issue:44936", + "right": "issue:45823", + "accept": false, + "reason": "Trainer lifecycle failure is unrelated to Gemma4 PLE device-map forward mismatch." + }, + { + "left": "issue:43653", + "right": "issue:44568", + "accept": false, + "reason": "BigBird special-token registration bug is unrelated to mdeberta-v3 BOS/EOS addition regression." + }, + { + "left": "issue:44464", + "right": "issue:44743", + "accept": false, + "reason": "Different generation bugs: compiled forward chunking versus Qwen3.5 recurrent state reset with cache." + }, + { + "left": "issue:38175", + "right": "issue:43295", + "accept": false, + "reason": "Siglip2 zero-probabilities issue is unrelated to processor.tokenizer regression." + }, + { + "left": "issue:43643", + "right": "issue:44561", + "accept": false, + "reason": "Both involve trust_remote_code, but one is missing AutoConfig fields and the other is v5 removal of is_torch_fx_available." + }, + { + "left": "issue:43653", + "right": "issue:44361", + "accept": false, + "reason": "Different tokenizer failures: BigBird mask-token registration versus MLukeTokenizer AttributeError." + }, + { + "left": "issue:44361", + "right": "issue:44442", + "accept": false, + "reason": "MLukeTokenizer task AttributeError is unrelated to FastSpeech2ConformerTokenizer loading." + }, + { + "left": "issue:30064", + "right": "issue:33666", + "accept": false, + "reason": "Void segmentation map processing and Qwen2-VL multi-GPU training are different problems." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet, and the only accepted soft-edge pair is a conservative, plausible tied-weights duplicate family. The other verdicts stay appropriately cautious and do not overstate shared code paths or bug identity." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 20, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:28282", + "issue:32090", + "issue:33357", + "issue:33666", + "issue:34567", + "issue:35141", + "issue:36010", + "issue:41720", + "issue:43122", + "issue:43278", + "issue:43452", + "issue:43493", + "issue:43504", + "issue:43525", + "issue:43531", + "issue:43550", + "issue:43572", + "issue:43577", + "issue:43606", + "issue:43646", + "issue:43653", + "issue:43688", + "issue:43697", + "issue:43720", + "issue:43742", + "issue:43749", + "issue:43825", + "issue:43906", + "issue:43927", + "issue:44361", + "issue:44451", + "issue:44464", + "issue:44479", + "issue:44554", + "issue:44568", + "issue:44610", + "issue:44625", + "issue:44821", + "issue:44843", + "issue:44849", + "issue:44977", + "issue:44991", + "issue:45357", + "issue:45440", + "issue:45588", + "issue:45701", + "issue:45704", + "issue:45812" + ], + "soft_pairs": [ + "issue:41720|issue:43906", + "issue:43688|issue:45440", + "issue:43688|issue:45704", + "issue:43577|issue:43697", + "issue:28282|issue:32090", + "issue:28282|issue:33357", + "issue:28282|issue:35141", + "issue:43646|issue:43749", + "issue:44843|issue:44991", + "issue:43688|issue:45701", + "issue:43653|issue:43927", + "issue:43531|issue:43742", + "issue:43550|issue:43927", + "issue:44464|issue:44977", + "issue:44479|issue:44625", + "issue:43550|issue:44361", + "issue:43577|issue:44554", + "issue:43577|issue:45588", + "issue:43122|issue:43906", + "issue:43504|issue:43550", + "issue:43504|issue:43720", + "issue:32090|issue:34567", + "issue:41720|issue:44464", + "issue:43504|issue:44610", + "issue:32090|issue:33666", + "issue:43452|issue:44451", + "issue:44849|issue:45357", + "issue:43742|issue:44821", + "issue:43278|issue:43688", + "issue:43493|issue:43825", + "issue:33666|issue:36010", + "issue:43606|issue:43927", + "issue:43720|issue:43927", + "issue:44568|issue:45812", + "issue:43525|issue:43572" + ], + "prepared_review_unit_hash": "2b5359add63ec35c01e6822a4d9bf03bbcedf01e415c3393aeb2b1ad78d477c4", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 35, + "serialized_chars": 38168, + "estimated_input_tokens": 9542, + "estimated_eval_tokens": 19340 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This is a heterogeneous cluster spanning unrelated tokenizer, model-loading, training, generation, and config bugs. None of the proposed soft pairs look like the same underlying bug/change, so all soft edges are rejected.", + "confidence": 0.95, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": "issue:45812 is the strongest standalone representative: it is a concrete, reproducible tokenizer regression with broad impact across Granite models and clear v4\u2192v5 behavior change.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:41720", + "right": "issue:43906", + "accept": false, + "reason": "Different failures: auto device mapping CUDA assert on A800 vs an isolated reproduction of another issue." + }, + { + "left": "issue:43688", + "right": "issue:45440", + "accept": false, + "reason": "Different bugs and code paths: auxiliary-loss normalization vs native DeepSeekV3 divergence." + }, + { + "left": "issue:43688", + "right": "issue:45704", + "accept": false, + "reason": "Different model/code paths: expert-loss normalization vs T5 apex FusedRMSNorm memory leak." + }, + { + "left": "issue:43577", + "right": "issue:43697", + "accept": false, + "reason": "Different symptoms and models: Blip2 dtype remaining fp32 vs RTDetrV2 output differences." + }, + { + "left": "issue:28282", + "right": "issue:32090", + "accept": false, + "reason": "ImportError for missing PyTorch vs Trainer GPU broadcast NoneType TypeError are unrelated." + }, + { + "left": "issue:28282", + "right": "issue:33357", + "accept": false, + "reason": "Missing PyTorch ImportError vs MacOS CLIP bus error are unrelated failures." + }, + { + "left": "issue:28282", + "right": "issue:35141", + "accept": false, + "reason": "Different bug classes: dependency import failure vs embedding resize reinitialization." + }, + { + "left": "issue:43646", + "right": "issue:43749", + "accept": false, + "reason": "Both are v5/loading-related, but they target different mechanisms: custom init breakage vs FSDP CPU RAM efficient loading." + }, + { + "left": "issue:44843", + "right": "issue:44991", + "accept": false, + "reason": "Different tokenizer paths: unconditional model_info call in Mistral regex patch vs tokenizer loading for EMBEDDIA/est-roberta." + }, + { + "left": "issue:43688", + "right": "issue:45701", + "accept": false, + "reason": "Auxiliary-loss normalization and tokenization changes are unrelated problems." + }, + { + "left": "issue:43653", + "right": "issue:43927", + "accept": false, + "reason": "BigBirdTokenizer special-token registration vs DiaConfig save/load token-ID loss are different bugs." + }, + { + "left": "issue:43531", + "right": "issue:43742", + "accept": false, + "reason": "Qwen3-MoE sliding_window issue and MobileLLM load key error are different code paths." + }, + { + "left": "issue:43550", + "right": "issue:43927", + "accept": false, + "reason": "torch.compile/SDPA behavior in Bamba is unrelated to DiaConfig token-ID persistence." + }, + { + "left": "issue:44464", + "right": "issue:44977", + "accept": false, + "reason": "Chunked generation with compiled forward vs Qwen3.5 flash-attention generation failure are different issues." + }, + { + "left": "issue:44479", + "right": "issue:44625", + "accept": false, + "reason": "Video-input regression for Qwen VL models vs num_labels propagation in Qwen3.5 are unrelated." + }, + { + "left": "issue:43550", + "right": "issue:44361", + "accept": false, + "reason": "Bamba compile/SDPA bug vs MLukeTokenizer AttributeError are unrelated." + }, + { + "left": "issue:43577", + "right": "issue:44554", + "accept": false, + "reason": "Blip2 dtype handling and an MPS attention correctness bug are different problems." + }, + { + "left": "issue:43577", + "right": "issue:45588", + "accept": false, + "reason": "Blip2 dtype retention vs flash_attention.py crashing on s_aux=None are unrelated." + }, + { + "left": "issue:43122", + "right": "issue:43906", + "accept": false, + "reason": "Different tokenization regression reports with no clear shared code path or concrete duplicate evidence." + }, + { + "left": "issue:43504", + "right": "issue:43550", + "accept": false, + "reason": "Beit legacy-field loading failure vs Bamba torch.compile/SDPA bug are unrelated." + }, + { + "left": "issue:43504", + "right": "issue:43720", + "accept": false, + "reason": "Beit preset loading and BitNet packed-weight unpacking are different load-time bugs." + }, + { + "left": "issue:32090", + "right": "issue:34567", + "accept": false, + "reason": "Trainer broadcast NoneType error vs num_input_tokens_seen not updating are separate Trainer issues." + }, + { + "left": "issue:41720", + "right": "issue:44464", + "accept": false, + "reason": "CUDA assert during auto device mapping is unrelated to chunked generation inconsistency." + }, + { + "left": "issue:43504", + "right": "issue:44610", + "accept": false, + "reason": "Legacy config field loading vs processor/model input-size mismatch are different bugs." + }, + { + "left": "issue:32090", + "right": "issue:33666", + "accept": false, + "reason": "Trainer broadcast NoneType error and Qwen2-VL multi-GPU training are not the same failure." + }, + { + "left": "issue:43452", + "right": "issue:44451", + "accept": false, + "reason": "GGUF/from_pretrained breakage and a specific model-load failure for ScandiBERT are different issues." + }, + { + "left": "issue:44849", + "right": "issue:45357", + "accept": false, + "reason": "Qwen3.5 hidden-states bug vs save_pretrained visual-encoder key regression are unrelated." + }, + { + "left": "issue:43742", + "right": "issue:44821", + "accept": false, + "reason": "MobileLLM key error vs AutoImageProcessor URL loading failure are different loading paths." + }, + { + "left": "issue:43278", + "right": "issue:43688", + "accept": false, + "reason": "Embedding dtype drift in evaluate vs auxiliary-loss normalization are unrelated." + }, + { + "left": "issue:43493", + "right": "issue:43825", + "accept": false, + "reason": "SigLIP2 implementation discrepancy vs pipeline translation error-message regression are different concerns." + }, + { + "left": "issue:33666", + "right": "issue:36010", + "accept": false, + "reason": "Qwen2-VL multi-GPU training and GenerationMixin import failure do not share the same bug." + }, + { + "left": "issue:43606", + "right": "issue:43927", + "accept": false, + "reason": "CPU-offload device mismatch for bark-small vs DiaConfig token-ID loss are unrelated." + }, + { + "left": "issue:43720", + "right": "issue:43927", + "accept": false, + "reason": "BitNet accelerate-loading unpacking bug vs DiaConfig token-ID persistence bug are different." + }, + { + "left": "issue:44568", + "right": "issue:45812", + "accept": false, + "reason": "Both are tokenizer regressions, but they affect different models and code paths: special-token injection for mdeberta-v3 vs wrong token IDs for Granite." + }, + { + "left": "issue:43525", + "right": "issue:43572", + "accept": false, + "reason": "Missing pad_token_id in Llama4Config and missing pad_token_idx in StableLmConfig are similar in shape but not the same concrete bug or fix." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 21, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:28218", + "issue:33357", + "issue:33666", + "issue:36331", + "issue:39692", + "issue:43257", + "issue:43262", + "issue:43278", + "issue:43295", + "issue:43299", + "issue:43334", + "issue:43493", + "issue:43504", + "issue:43540", + "issue:43572", + "issue:43575", + "issue:43606", + "issue:43646", + "issue:43653", + "issue:43720", + "issue:43756", + "issue:43824", + "issue:43844", + "issue:43927", + "issue:43994", + "issue:44112", + "issue:44291", + "issue:44361", + "issue:44423", + "issue:44464", + "issue:44466", + "issue:44488", + "issue:44493", + "issue:44589", + "issue:44625", + "issue:44734", + "issue:44843", + "issue:44913", + "issue:44964", + "issue:44977", + "issue:45083", + "issue:45200", + "issue:45229", + "issue:45405", + "issue:45520", + "issue:45584", + "issue:45759", + "issue:45823" + ], + "soft_pairs": [ + "issue:43262|issue:43295", + "issue:43756|issue:44913", + "issue:43720|issue:44361", + "issue:28218|issue:33357", + "issue:33666|issue:36331", + "issue:39692|issue:43493", + "issue:43540|issue:44977", + "issue:43493|issue:43844", + "issue:44964|issue:45405", + "issue:43278|issue:43844", + "issue:43334|issue:43572", + "issue:44625|issue:45584", + "issue:43257|issue:43299", + "issue:43756|issue:45083", + "issue:43334|issue:45759", + "issue:44488|issue:44843", + "issue:45200|issue:45229", + "issue:43824|issue:44589", + "issue:43646|issue:44466", + "issue:39692|issue:43295", + "issue:45229|issue:45823", + "issue:43504|issue:43606", + "issue:44589|issue:45520", + "issue:43572|issue:44291", + "issue:33357|issue:33666", + "issue:44423|issue:44734", + "issue:43927|issue:44361", + "issue:43504|issue:43653", + "issue:44464|issue:45584", + "issue:43606|issue:44112", + "issue:43994|issue:44493", + "issue:43575|issue:45405" + ], + "prepared_review_unit_hash": "9fff9d47fd7ccaa3e4cfbc8ef9993c69c2f175e6b0b20c345c2c816c5b112857", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 32, + "serialized_chars": 37543, + "estimated_input_tokens": 9386, + "estimated_eval_tokens": 19028 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is heterogeneous and the soft pairs are only broad thematic matches, not true duplicates. I would reject all proposed duplicate edges and avoid selecting a canonical issue/PR for the cluster.", + "confidence": 0.95, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43262", + "right": "issue:43295", + "accept": false, + "reason": "Audio sample-rate default bug vs processor/tokenizer regression; different subsystems and failure modes." + }, + { + "left": "issue:43756", + "right": "issue:44913", + "accept": false, + "reason": "Both mention rotary settings, but one is a Smollm3 RoPE-layer implementation issue and the other is a GPTNeoX config reload bug." + }, + { + "left": "issue:43720", + "right": "issue:44361", + "accept": false, + "reason": "BitNet packed-weight loading failure vs MLukeTokenizer task-time AttributeError; unrelated code paths." + }, + { + "left": "issue:28218", + "right": "issue:33357", + "accept": false, + "reason": "Tokenizer spacing bug vs MacOS bus error on a CLIP model; not the same defect." + }, + { + "left": "issue:33666", + "right": "issue:36331", + "accept": false, + "reason": "Qwen2-VL multi-GPU training request vs custom trainer loss-signature TypeError; different problems." + }, + { + "left": "issue:39692", + "right": "issue:43493", + "accept": false, + "reason": "Documentation/example errors vs SigLIP2 implementation parity discrepancy; not the same underlying bug." + }, + { + "left": "issue:43540", + "right": "issue:44977", + "accept": false, + "reason": "Qwen3OmniMoe video-input processing vs Qwen3.5 flash-attention generation issue; different code paths." + }, + { + "left": "issue:43493", + "right": "issue:43844", + "accept": false, + "reason": "SigLIP2 model mismatch issue vs DeepSpeed/gradient behavior regression; unrelated." + }, + { + "left": "issue:44964", + "right": "issue:45405", + "accept": false, + "reason": "Phi-4 multimodal loading failure vs PEFT version pin bump; no shared bug." + }, + { + "left": "issue:43278", + "right": "issue:43844", + "accept": false, + "reason": "Eval/train dtype mismatch vs ZeRO-3 gradient inflation; different defects." + }, + { + "left": "issue:43334", + "right": "issue:43572", + "accept": false, + "reason": "Both involve missing pad-token fields, but in different models/config layers and not the same concrete bug." + }, + { + "left": "issue:44625", + "right": "issue:45584", + "accept": false, + "reason": "Qwen3.5 num_labels propagation bug vs Whisper empty-transcription generation failure; unrelated." + }, + { + "left": "issue:43257", + "right": "issue:43299", + "accept": false, + "reason": "Both are Qwen MOE-related, but one is accelerate/deepspeed weight conversion and the other is multimodal loading; not mergeable as one fix." + }, + { + "left": "issue:43756", + "right": "issue:45083", + "accept": false, + "reason": "Smollm3 RoPE-layer mismatch vs qwen3_omni_moe feature-length helper behavior; different bugs." + }, + { + "left": "issue:43334", + "right": "issue:45759", + "accept": false, + "reason": "Missing pad_token_id attribute vs unwrap text_config in from_config; related family, but distinct code paths." + }, + { + "left": "issue:44488", + "right": "issue:44843", + "accept": false, + "reason": "Model load failure for a specific model vs offline-mode model_info call; unrelated." + }, + { + "left": "issue:45200", + "right": "issue:45229", + "accept": false, + "reason": "Text-only token-type defaulting vs multi-GPU inference OOM; different issues." + }, + { + "left": "issue:43824", + "right": "issue:44589", + "accept": false, + "reason": "ImportError for a Qwen2.5-VL class vs missing Float8 storage type; unrelated failures." + }, + { + "left": "issue:43646", + "right": "issue:44466", + "accept": false, + "reason": "Custom model initialization regression vs device-dependent serialization inconsistency; not the same bug." + }, + { + "left": "issue:39692", + "right": "issue:43295", + "accept": false, + "reason": "SigLIP2 docs/example problems vs processor/tokenizer regression; different scopes and code paths." + }, + { + "left": "issue:45229", + "right": "issue:45823", + "accept": false, + "reason": "Both Gemma4-related, but one is CUDA OOM and the other is a device mismatch during forward; not the same defect." + }, + { + "left": "issue:43504", + "right": "issue:43606", + "accept": false, + "reason": "Beit legacy-field loading bug vs Bark CPU-offload device mismatch; unrelated." + }, + { + "left": "issue:44589", + "right": "issue:45520", + "accept": false, + "reason": "Float8 storage lookup failure vs flash_attn import-utils KeyError on Python 3.13; different issues." + }, + { + "left": "issue:43572", + "right": "issue:44291", + "accept": false, + "reason": "Both are v5 init/config regressions, but one is a StableLmConfig field issue and the other is an init_empty_weights argument change." + }, + { + "left": "issue:33357", + "right": "issue:33666", + "accept": false, + "reason": "MacOS CLIP bus error vs Qwen2-VL multi-GPU training; no common bug." + }, + { + "left": "issue:44423", + "right": "issue:44734", + "accept": false, + "reason": "Serve crash on multimodal model input conversion vs KV-cache continuation tensor indexing; different serve paths." + }, + { + "left": "issue:43927", + "right": "issue:44361", + "accept": false, + "reason": "DiaConfig save/load token-ID loss vs MLukeTokenizer AttributeError; unrelated." + }, + { + "left": "issue:43504", + "right": "issue:43653", + "accept": false, + "reason": "Legacy-field model load bug vs BigBirdTokenizer special-token registration issue; different fixes." + }, + { + "left": "issue:44464", + "right": "issue:45584", + "accept": false, + "reason": "Chunked generation inconsistency under compiled forward vs Whisper empty transcription after align_special_tokens; unrelated." + }, + { + "left": "issue:43606", + "right": "issue:44112", + "accept": false, + "reason": "Real CPU-offload device mismatch bug vs a stale GraniteSpeech CI test; not duplicates." + }, + { + "left": "issue:43994", + "right": "issue:44493", + "accept": false, + "reason": "SigLIP2 pipeline output issue vs broad unexpected position-id keys; not the same underlying problem." + }, + { + "left": "issue:43575", + "right": "issue:45405", + "accept": false, + "reason": "Tensor-parallel OOM on a specific model vs unreleased PEFT version bump; unrelated." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 22, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:15354", + "issue:28218", + "issue:29942", + "issue:30333", + "issue:34567", + "issue:37428", + "issue:38175", + "issue:42490", + "issue:43232", + "issue:43278", + "issue:43295", + "issue:43493", + "issue:43540", + "issue:43572", + "issue:43582", + "issue:43696", + "issue:43701", + "issue:43824", + "issue:43844", + "issue:43873", + "issue:43957", + "issue:44315", + "issue:44464", + "issue:44466", + "issue:44521", + "issue:44554", + "issue:44610", + "issue:44792", + "issue:44849", + "issue:44898", + "issue:44936", + "issue:44964", + "issue:44977", + "issue:44987", + "issue:45072", + "issue:45084", + "issue:45127", + "issue:45198", + "issue:45200", + "issue:45290", + "issue:45325", + "issue:45356", + "issue:45405", + "issue:45412", + "issue:45446", + "issue:45520", + "issue:45563", + "issue:45823" + ], + "soft_pairs": [ + "issue:43278|issue:43701", + "issue:44792|issue:45405", + "issue:44521|issue:45290", + "issue:29942|issue:30333", + "issue:42490|issue:43696", + "issue:43844|issue:44964", + "issue:28218|issue:34567", + "issue:44554|issue:45446", + "issue:45084|issue:45520", + "issue:43582|issue:44977", + "issue:43540|issue:45200", + "issue:15354|issue:38175", + "issue:43232|issue:44849", + "issue:45072|issue:45198", + "issue:37428|issue:43957", + "issue:45200|issue:45823", + "issue:44315|issue:44849", + "issue:44554|issue:45356", + "issue:43493|issue:43701", + "issue:43295|issue:44610", + "issue:43873|issue:44464", + "issue:44898|issue:45412", + "issue:44466|issue:45127", + "issue:44792|issue:45563", + "issue:43572|issue:44987", + "issue:45200|issue:45325", + "issue:37428|issue:43824", + "issue:43572|issue:43957", + "issue:43844|issue:44936" + ], + "prepared_review_unit_hash": "eee1573adc68cd615b64d0c7f0f40d98a8056232111f73de13044540dfe69725", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 29, + "serialized_chars": 36425, + "estimated_input_tokens": 9107, + "estimated_eval_tokens": 18470 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is heterogeneous: the soft links mostly share broad areas like tokenization, generation, or multimodal models, but they do not describe the same concrete bug or change. No duplicate canonical issue/PR stands out.", + "confidence": 0.94, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43278", + "right": "issue:43701", + "accept": false, + "reason": "BF16->FP32 dtype drift in evaluation and resume_from_checkpoint key mismatch are different failures with different code paths." + }, + { + "left": "issue:44792", + "right": "issue:45405", + "accept": false, + "reason": "A Janus model test failure and an unreleased PEFT version bump are unrelated." + }, + { + "left": "issue:44521", + "right": "issue:45290", + "accept": false, + "reason": "Both involve chat templating, but one is wrong assistant masks for multimodal inputs and the other is a crash on tool-call assistant messages; different bugs." + }, + { + "left": "issue:29942", + "right": "issue:30333", + "accept": false, + "reason": "Flash Attention 2 test failures and MLflow reporting after job failure are unrelated." + }, + { + "left": "issue:42490", + "right": "issue:43696", + "accept": false, + "reason": "Local-path import warnings and CUDA OOM loading GPT-oss-20b are not the same issue." + }, + { + "left": "issue:43844", + "right": "issue:44964", + "accept": false, + "reason": "ZeRO-3 gradient growth and a model load/import failure are different problems." + }, + { + "left": "issue:28218", + "right": "issue:34567", + "accept": false, + "reason": "Tokenizer spacing after added tokens and TrainerState token counters not updating are separate behaviors." + }, + { + "left": "issue:44554", + "right": "issue:45446", + "accept": false, + "reason": "An MPS attention correctness bug and a PyTorch version check for AuxRequest import are unrelated." + }, + { + "left": "issue:45084", + "right": "issue:45520", + "accept": false, + "reason": "Template-node compilation errors and a flash_attn import KeyError on Python 3.13 are distinct import/compile paths." + }, + { + "left": "issue:43582", + "right": "issue:44977", + "accept": false, + "reason": "Apple Silicon warmup TypeError and Qwen3.5 flash-attention generation issues are not the same bug." + }, + { + "left": "issue:43540", + "right": "issue:45200", + "accept": false, + "reason": "Video-processing ValueError in Qwen3OmniMoe and Gemma4 text-only mm_token_type_ids defaults are different multimodal issues." + }, + { + "left": "issue:15354", + "right": "issue:38175", + "accept": false, + "reason": "TorchScript GeneratorExp export failure and SigLIP2 zero probabilities are unrelated model bugs." + }, + { + "left": "issue:43232", + "right": "issue:44849", + "accept": false, + "reason": "Generation kwargs after sync_gpus and a Qwen3.5 hidden-states bug do not share the same underlying defect." + }, + { + "left": "issue:45072", + "right": "issue:45198", + "accept": false, + "reason": "bfloat16 dtype mismatches in inference and wav2vec2 save/tokenization failures are different code paths." + }, + { + "left": "issue:37428", + "right": "issue:43957", + "accept": false, + "reason": "Missing flash_supports_window_size import and meta-device loading regressions are separate import/model-loading bugs." + }, + { + "left": "issue:45200", + "right": "issue:45823", + "accept": false, + "reason": "Gemma4 mm_token_type_ids defaults and device-map forward mismatches are not the same problem." + }, + { + "left": "issue:44315", + "right": "issue:44849", + "accept": false, + "reason": "Liger Kernel application with model_init and Qwen3.5 hidden-state behavior are unrelated." + }, + { + "left": "issue:44554", + "right": "issue:45356", + "accept": false, + "reason": "MPS attention correctness and Kimi-K2.5 tokenizer regression are different subsystems and failures." + }, + { + "left": "issue:43493", + "right": "issue:43701", + "accept": false, + "reason": "SigLIP2 HF-vs-JAX discrepancy and checkpoint resume key mismatch are unrelated." + }, + { + "left": "issue:43295", + "right": "issue:44610", + "accept": false, + "reason": "Processor/tokenizer regression and OmDet-Turbo input-size mismatch are different bugs." + }, + { + "left": "issue:43873", + "right": "issue:44464", + "accept": false, + "reason": "Quantization offloading behavior and chunked generation under compiled forward are not the same issue." + }, + { + "left": "issue:44898", + "right": "issue:45412", + "accept": false, + "reason": "Perceiver resolution interpolation failure and RT-DETR memory release leak are unrelated." + }, + { + "left": "issue:44466", + "right": "issue:45127", + "accept": false, + "reason": "lm_head.weight serialization inconsistency and LoRA merge collapse with extended vocab are different failure modes." + }, + { + "left": "issue:44792", + "right": "issue:45563", + "accept": false, + "reason": "A Janus model test failure and a stale warning in paged generate() are not the same bug." + }, + { + "left": "issue:43572", + "right": "issue:44987", + "accept": false, + "reason": "Missing pad_token_idx in config and a transformers>=5.1.0 load failure are only loosely related versioned-model issues, not one concrete defect." + }, + { + "left": "issue:45200", + "right": "issue:45325", + "accept": false, + "reason": "Gemma4 token-type defaults and Qwen2.5-VL rope index scaling are different multimodal bugs." + }, + { + "left": "issue:37428", + "right": "issue:43824", + "accept": false, + "reason": "These are both import errors, but they concern different missing symbols for different models." + }, + { + "left": "issue:43572", + "right": "issue:43957", + "accept": false, + "reason": "A missing config field and a meta-device loading regression are not the same underlying change." + }, + { + "left": "issue:43844", + "right": "issue:44936", + "accept": false, + "reason": "ZeRO-3 gradient instability and trainer.evaluate() failing after train() are different training problems." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 23, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:28218", + "issue:35141", + "issue:37428", + "issue:38175", + "issue:42490", + "issue:42915", + "issue:43262", + "issue:43278", + "issue:43295", + "issue:43334", + "issue:43335", + "issue:43388", + "issue:43425", + "issue:43450", + "issue:43493", + "issue:43525", + "issue:43526", + "issue:43550", + "issue:43575", + "issue:43606", + "issue:43653", + "issue:43819", + "issue:43824", + "issue:43825", + "issue:43827", + "issue:43856", + "issue:43873", + "issue:43901", + "issue:43975", + "issue:44112", + "issue:44206", + "issue:44442", + "issue:44448", + "issue:44514", + "issue:44521", + "issue:44610", + "issue:44743", + "issue:44857", + "issue:45083", + "issue:45127", + "issue:45198", + "issue:45245", + "issue:45290", + "issue:45381", + "issue:45412", + "issue:45584", + "issue:45593" + ], + "soft_pairs": [ + "issue:43295|issue:43450", + "issue:43334|issue:43525", + "issue:43824|issue:43975", + "issue:43295|issue:43873", + "issue:43388|issue:43526", + "issue:38175|issue:45412", + "issue:42915|issue:43824", + "issue:44514|issue:44521", + "issue:38175|issue:43493", + "issue:44442|issue:45198", + "issue:28218|issue:35141", + "issue:43278|issue:43856", + "issue:42490|issue:43493", + "issue:43653|issue:44112", + "issue:43827|issue:43901", + "issue:43450|issue:43575", + "issue:44206|issue:45584", + "issue:45127|issue:45245", + "issue:43819|issue:45083", + "issue:43819|issue:44112", + "issue:43262|issue:43450", + "issue:44610|issue:44857", + "issue:43278|issue:43825", + "issue:37428|issue:43425", + "issue:43606|issue:45083", + "issue:45290|issue:45381", + "issue:43335|issue:43550", + "issue:44448|issue:44743", + "issue:44857|issue:45593" + ], + "prepared_review_unit_hash": "dd08ec80d0a6a3130b7b16dd0883e6c98f47312e92c9838e41a27d80e6178b7d", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 29, + "serialized_chars": 35995, + "estimated_input_tokens": 8999, + "estimated_eval_tokens": 18254 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This set is mostly heterogeneous: tokenization, multimodal chat-template, import, runtime, and training issues that only look superficially similar. I found one plausible duplicate family around missing `pad_token_id` on config objects; the rest should stay separate.", + "confidence": 0.81, + "canonical_issue_reason": "issue:43334 is the clearest anchor for the only plausible duplicate family here: an `AttributeError` from missing `pad_token_id` on config objects, with a near-identical sibling in issue:43525.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43334 is the best overall issue to anchor the cluster because it has a concrete, model-loading code-path bug and a plausible duplicate; the other items are mostly unrelated or only broadly similar.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43295", + "right": "issue:43450", + "accept": false, + "reason": "Different regressions: processor/tokenizer API breakage vs batched video shape handling." + }, + { + "left": "issue:43334", + "right": "issue:43525", + "accept": true, + "reason": "Both report the same underlying `pad_token_id` config attribute bug causing load-time `AttributeError`." + }, + { + "left": "issue:43824", + "right": "issue:43975", + "accept": false, + "reason": "Unrelated failures: an import error for a missing class vs a detokenization/output issue." + }, + { + "left": "issue:43295", + "right": "issue:43873", + "accept": false, + "reason": "Processor regression and quantization/offloading behavior are different code paths and symptoms." + }, + { + "left": "issue:43388", + "right": "issue:43526", + "accept": false, + "reason": "Different bugs in different helpers: metric gathering label truncation vs image processor label reduction." + }, + { + "left": "issue:38175", + "right": "issue:45412", + "accept": false, + "reason": "SigLIP2 output anomaly and RT-DETR memory retention are unrelated." + }, + { + "left": "issue:42915", + "right": "issue:43824", + "accept": false, + "reason": "FP8 config failure and Qwen2.5-VL import failure are different issues." + }, + { + "left": "issue:44514", + "right": "issue:44521", + "accept": false, + "reason": "Both involve `apply_chat_template`, but one crashes on batched input and the other produces incorrect assistant masks; not the same bug." + }, + { + "left": "issue:38175", + "right": "issue:43493", + "accept": false, + "reason": "Same model family, but the titles point to different symptoms and not clearly the same fix." + }, + { + "left": "issue:44442", + "right": "issue:45198", + "accept": false, + "reason": "Tokenizer load failure vs save/tokenization failure in a different model family." + }, + { + "left": "issue:28218", + "right": "issue:35141", + "accept": false, + "reason": "Tokenizer spacing behavior and embedding reinitialization are unrelated." + }, + { + "left": "issue:43278", + "right": "issue:43856", + "accept": false, + "reason": "Training/evaluation dtype change and MoE memory usage are distinct problems." + }, + { + "left": "issue:42490", + "right": "issue:43493", + "accept": false, + "reason": "Local-path import warning and SigLIP2 implementation discrepancy are unrelated." + }, + { + "left": "issue:43653", + "right": "issue:44112", + "accept": false, + "reason": "Tokenizer special-token registration bug vs a stale CI device-override test." + }, + { + "left": "issue:43827", + "right": "issue:43901", + "accept": false, + "reason": "Both are docs-related, but they cover different API deprecations and different pages." + }, + { + "left": "issue:43450", + "right": "issue:43575", + "accept": false, + "reason": "Batched video shape bug and tensor-parallel OOM are unrelated." + }, + { + "left": "issue:44206", + "right": "issue:45584", + "accept": false, + "reason": "Feature-extractor center-arg crash and Whisper empty-transcription handling are different failures." + }, + { + "left": "issue:45127", + "right": "issue:45245", + "accept": false, + "reason": "LoRA/vocabulary merge collapse and category-count overflow are not the same bug." + }, + { + "left": "issue:43819", + "right": "issue:45083", + "accept": false, + "reason": "DAC STE mismatch and qwen3_omni_moe length helper behavior are different code paths." + }, + { + "left": "issue:43819", + "right": "issue:44112", + "accept": false, + "reason": "DAC forward mismatch and a GraniteSpeech CI test issue are unrelated." + }, + { + "left": "issue:43262", + "right": "issue:43450", + "accept": false, + "reason": "Audio chat-template sample-rate defaulting and video batching shape are different subsystems." + }, + { + "left": "issue:44610", + "right": "issue:44857", + "accept": false, + "reason": "Processor input-size mismatch and AMP/CUDA loss crash are separate bugs." + }, + { + "left": "issue:43278", + "right": "issue:43825", + "accept": false, + "reason": "Embedding dtype regression and pipeline docs mismatch are unrelated." + }, + { + "left": "issue:37428", + "right": "issue:43425", + "accept": false, + "reason": "A specific import error and a Torch version incompatibility are not the same issue." + }, + { + "left": "issue:43606", + "right": "issue:45083", + "accept": false, + "reason": "CPU offload device mismatch and qwen3_omni_moe feature-length helper behavior are unrelated." + }, + { + "left": "issue:45290", + "right": "issue:45381", + "accept": false, + "reason": "Chat-template tool-call crash and video vision-position-id bug are different multimodal issues." + }, + { + "left": "issue:43335", + "right": "issue:43550", + "accept": false, + "reason": "Sparse-layer construction bug and Bamba SDPA/compile failure do not share the same root cause." + }, + { + "left": "issue:44448", + "right": "issue:44743", + "accept": false, + "reason": "Output mismatch for Pegasus and recurrent-state reset in modular Qwen3 are unrelated." + }, + { + "left": "issue:44857", + "right": "issue:45593", + "accept": false, + "reason": "AMP/CUDA crash and missing auxiliary losses are different model-training bugs." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The analysis is grounded in the packet and stays conservative. The anchor choice around issue:43334 is supported by the matching `pad_token_id` AttributeError in issue:43525, and the summary does not overclaim broader duplication beyond that plausible family. The other pairwise rejections are phrased cautiously and are consistent with the issue titles." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 24, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:29127", + "issue:29942", + "issue:32090", + "issue:34689", + "issue:35141", + "issue:37428", + "issue:38175", + "issue:39692", + "issue:40444", + "issue:42490", + "issue:43335", + "issue:43504", + "issue:43526", + "issue:43550", + "issue:43606", + "issue:43611", + "issue:43653", + "issue:43720", + "issue:43723", + "issue:43819", + "issue:43854", + "issue:43931", + "issue:44112", + "issue:44206", + "issue:44292", + "issue:44351", + "issue:44360", + "issue:44448", + "issue:44462", + "issue:44493", + "issue:44554", + "issue:44561", + "issue:44779", + "issue:44855", + "issue:44857", + "issue:44977", + "issue:45072", + "issue:45083", + "issue:45200", + "issue:45229", + "issue:45310", + "issue:45478", + "issue:45520", + "issue:45563", + "issue:45588", + "issue:45593", + "issue:45823", + "issue:45859" + ], + "soft_pairs": [ + "issue:44360|issue:44561", + "issue:45200|issue:45588", + "issue:37428|issue:44351", + "issue:29942|issue:32090", + "issue:44206|issue:45823", + "issue:43611|issue:43931", + "issue:39692|issue:43526", + "issue:43335|issue:43653", + "issue:43819|issue:44493", + "issue:43550|issue:44112", + "issue:43335|issue:43504", + "issue:44855|issue:45083", + "issue:45310|issue:45478", + "issue:43335|issue:45083", + "issue:29127|issue:32090", + "issue:45588|issue:45859", + "issue:39692|issue:44493", + "issue:44292|issue:45520", + "issue:38175|issue:44493", + "issue:43335|issue:45072", + "issue:44448|issue:44779", + "issue:44462|issue:44561", + "issue:44561|issue:44779", + "issue:43335|issue:45593", + "issue:43720|issue:44112", + "issue:44977|issue:45229", + "issue:42490|issue:43854", + "issue:40444|issue:43526", + "issue:44448|issue:44977", + "issue:34689|issue:35141", + "issue:43606|issue:44857", + "issue:43723|issue:44554", + "issue:44857|issue:45072", + "issue:44977|issue:45563" + ], + "prepared_review_unit_hash": "2f25421e8c38744aae2da9fb5ff9b1f023f6ad5de640618830760843e9e5fc3c", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 34, + "serialized_chars": 37895, + "estimated_input_tokens": 9474, + "estimated_eval_tokens": 19204 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Only one soft edge looks like a true duplicate pair; the rest share keywords but describe different bugs, regressions, or subsystems.", + "confidence": 0.97, + "canonical_issue_reason": "issue:45310 is the earlier, fuller report of the same Qwen3.5 MoE `from_pretrained` regression as issue:45478.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45310 is the clearest duplicate target in this set: specific, reproducible, and matched by the near-identical follow-up report.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44360", + "right": "issue:44561", + "accept": false, + "reason": "Different code paths: DSA indexer/ReLU discussion vs v5 trust_remote_code breakage from `is_torch_fx_available` removal." + }, + { + "left": "issue:45200", + "right": "issue:45588", + "accept": false, + "reason": "Different failures: missing `mm_token_type_ids` for Gemma 4 fine-tuning vs flash-attention crash on `s_aux=None`." + }, + { + "left": "issue:37428", + "right": "issue:44351", + "accept": false, + "reason": "Both are import errors, but for different missing symbols (`_flash_supports_window_size` vs `HybridCache`)." + }, + { + "left": "issue:29942", + "right": "issue:32090", + "accept": false, + "reason": "Flash Attention 2 test failures are unrelated to the Trainer `_gpu_broadcast_one` NoneType broadcast error." + }, + { + "left": "issue:44206", + "right": "issue:45823", + "accept": false, + "reason": "Different subsystems and symptoms: feature extractor `center` arg crash vs Gemma4 device-map mismatch during forward." + }, + { + "left": "issue:43611", + "right": "issue:43931", + "accept": false, + "reason": "Model-loading regression from `base_model_prefix` handling is distinct from Qwen3-VL weight-shape mismatch." + }, + { + "left": "issue:39692", + "right": "issue:43526", + "accept": false, + "reason": "SigLIP2 docs/example issues are unrelated to BeitImageProcessorFast `reduce_labels` returning one label." + }, + { + "left": "issue:43335", + "right": "issue:43653", + "accept": false, + "reason": "SwitchTransformers sparse-layer config bug is unrelated to BigBirdTokenizer special-token decode behavior." + }, + { + "left": "issue:43819", + "right": "issue:44493", + "accept": false, + "reason": "DAC `from_latents`/STE mismatch is a different bug from unexpected `position_id` keys across models." + }, + { + "left": "issue:43550", + "right": "issue:44112", + "accept": false, + "reason": "Different targets: Bamba torch.compile+SDPA failure vs stale GraniteSpeech CI test/device override issue." + }, + { + "left": "issue:43335", + "right": "issue:43504", + "accept": false, + "reason": "Same broad model family, but different bugs: sparse-layer creation vs Beit legacy-field loading preset failure." + }, + { + "left": "issue:44855", + "right": "issue:45083", + "accept": false, + "reason": "Python 3.13 JIT parse/indentation import error is unrelated to qwen3_omni_moe feature-length helper behavior." + }, + { + "left": "issue:45310", + "right": "issue:45478", + "accept": true, + "reason": "Near-identical reports of the same Qwen3.5 MoE `from_pretrained` regression on transformers>=5.4.0." + }, + { + "left": "issue:43335", + "right": "issue:45083", + "accept": false, + "reason": "SwitchTransformers config bug and qwen3_omni_moe feature-length helper bug are unrelated." + }, + { + "left": "issue:29127", + "right": "issue:32090", + "accept": false, + "reason": "LayoutLMV3 box/clarity error is unrelated to Trainer `_gpu_broadcast_one` NoneType broadcast failure." + }, + { + "left": "issue:45588", + "right": "issue:45859", + "accept": false, + "reason": "Different flash-attention edge cases: sink-less `s_aux=None` crash vs CPU-offload incompatibility in Qwen3_5MoE rotary embedding." + }, + { + "left": "issue:39692", + "right": "issue:44493", + "accept": false, + "reason": "SigLIP2 documentation/model mismatch is unrelated to unexpected `position_id` keys." + }, + { + "left": "issue:44292", + "right": "issue:45520", + "accept": false, + "reason": "Qwen-3-8B-NVFP4 runtime failure is unrelated to the Python 3.13 `flash_attn` import key error." + }, + { + "left": "issue:38175", + "right": "issue:44493", + "accept": false, + "reason": "Zero-probability outputs in SigLIP2 are a different symptom from unexpected `position_id` key warnings." + }, + { + "left": "issue:43335", + "right": "issue:45072", + "accept": false, + "reason": "Different failure modes: SwitchTransformers sparse-layer config vs dtype mismatches in bfloat16 inference." + }, + { + "left": "issue:44448", + "right": "issue:44779", + "accept": false, + "reason": "Different regressions in different models: Pegasus v4/v5 output mismatch vs Deepseek tokenizer correctness change." + }, + { + "left": "issue:44462", + "right": "issue:44561", + "accept": false, + "reason": "AutoTokenizer ignoring `tokenizer.json` is unrelated to trust_remote_code breakage from removed `is_torch_fx_available`." + }, + { + "left": "issue:44561", + "right": "issue:44779", + "accept": false, + "reason": "Different areas: remote-code utility removal vs Deepseek tokenizer output regression." + }, + { + "left": "issue:43335", + "right": "issue:45593", + "accept": false, + "reason": "SwitchTransformers sparse-layer config bug is unrelated to D-FINE auxiliary-loss handling." + }, + { + "left": "issue:43720", + "right": "issue:44112", + "accept": false, + "reason": "BitNet accelerate-loading unpack bug is unrelated to GraniteSpeech stale CI device-override test." + }, + { + "left": "issue:44977", + "right": "issue:45229", + "accept": false, + "reason": "Qwen3.5 flash-attention generation bug is unrelated to Gemma4 multi-GPU CUDA OOM." + }, + { + "left": "issue:42490", + "right": "issue:43854", + "accept": false, + "reason": "Trailing-slash local model import warning is unrelated to GLM-4.7-Flash loading failure in tests." + }, + { + "left": "issue:40444", + "right": "issue:43526", + "accept": false, + "reason": "IterableDataset multi-image finetuning failure is unrelated to BeitImageProcessorFast `reduce_labels` bug." + }, + { + "left": "issue:44448", + "right": "issue:44977", + "accept": false, + "reason": "Different model/output regressions: Pegasus v4/v5 mismatch vs Qwen3.5 flash-attention generation issue." + }, + { + "left": "issue:34689", + "right": "issue:35141", + "accept": false, + "reason": "Llama 3.2 Vision loading breakage and token-embedding reinit on `post_init` are different bugs." + }, + { + "left": "issue:43606", + "right": "issue:44857", + "accept": false, + "reason": "CPU offload device mismatch in bark-small is unrelated to LwDetrImageLoss AMP/CUDA crash." + }, + { + "left": "issue:43723", + "right": "issue:44554", + "accept": false, + "reason": "Tokenizer loading in v5 and MPS attention correctness are different problems." + }, + { + "left": "issue:44857", + "right": "issue:45072", + "accept": false, + "reason": "Different failures: loss/AMP crash vs dtype mismatch in inference." + }, + { + "left": "issue:44977", + "right": "issue:45563", + "accept": false, + "reason": "Qwen3.5 flash-attention generation bug is unrelated to the stale `num_return_sequences` warning in paged generate()." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The canonical duplicate choice is well grounded: issue:45310 and issue:45478 have effectively the same title and clearly refer to the same Qwen3.5 MoE from_pretrained regression. The remaining soft-edge rejections are conservative and supported by the packet-level titles/symptoms, without overstating evidence." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 25, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:10000", + "issue:15354", + "issue:28218", + "issue:29127", + "issue:29942", + "issue:32090", + "issue:33357", + "issue:35141", + "issue:37428", + "issue:41720", + "issue:42490", + "issue:43065", + "issue:43278", + "issue:43329", + "issue:43381", + "issue:43412", + "issue:43550", + "issue:43575", + "issue:43653", + "issue:43701", + "issue:43873", + "issue:43927", + "issue:43976", + "issue:43994", + "issue:44448", + "issue:44466", + "issue:44493", + "issue:44568", + "issue:44743", + "issue:44857", + "issue:44987", + "issue:45083", + "issue:45198", + "issue:45229", + "issue:45245", + "issue:45372", + "issue:45406", + "issue:45412", + "issue:45440", + "issue:45459", + "issue:45491", + "issue:45563", + "issue:45588", + "issue:45698", + "issue:45712", + "issue:45803", + "issue:45823" + ], + "soft_pairs": [ + "issue:44568|issue:45491", + "issue:43927|issue:45698", + "issue:28218|issue:32090", + "issue:43701|issue:45083", + "issue:43550|issue:45083", + "issue:45198|issue:45803", + "issue:45372|issue:45588", + "issue:37428|issue:44987", + "issue:29942|issue:35141", + "issue:43873|issue:44493", + "issue:45459|issue:45712", + "issue:32090|issue:33357", + "issue:45245|issue:45563", + "issue:43412|issue:45588", + "issue:43873|issue:45229", + "issue:41720|issue:45229", + "issue:43278|issue:43329", + "issue:43976|issue:45440", + "issue:10000|issue:15354", + "issue:45372|issue:45406", + "issue:43065|issue:43994", + "issue:42490|issue:43575", + "issue:28218|issue:29127", + "issue:43065|issue:45412", + "issue:45372|issue:45459", + "issue:44857|issue:45588", + "issue:43653|issue:44743", + "issue:44448|issue:44466", + "issue:43278|issue:43381", + "issue:45588|issue:45823" + ], + "prepared_review_unit_hash": "8e9fcfb3eaf43d6f226da521c48ee5d0c3fc8045d73de8800f7e9f5a569d8b78", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 30, + "serialized_chars": 36365, + "estimated_input_tokens": 9092, + "estimated_eval_tokens": 18440 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is broadly heterogeneous. A few items share themes like tokenizer/save-load regressions or Gemma4 processor-loading failures, but none of the soft pairs look like the same concrete bug/change, so I reject all duplicate candidates.", + "confidence": 0.84, + "canonical_issue_reason": "issue:45372 is the most central representative: it is a current Gemma4 processor-loading regression, has downstream references, and sits closest to the small number of related Gemma4/processor items in the set.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45372 is the best single anchor for this set because it is specific, recent, open, and most likely to attract related follow-up reports.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44568", + "right": "issue:45491", + "accept": false, + "reason": "Tokenizer special-token behavior vs Gemma3 NaN embeddings are unrelated bugs." + }, + { + "left": "issue:43927", + "right": "issue:45698", + "accept": false, + "reason": "Both involve save/load flows, but one is config token IDs and the other is wrong custom module resolution." + }, + { + "left": "issue:28218", + "right": "issue:32090", + "accept": false, + "reason": "Tokenizer spacing after added tokens is unrelated to Trainer broadcast NoneType errors." + }, + { + "left": "issue:43701", + "right": "issue:45083", + "accept": false, + "reason": "Checkpoint key mismatch and multimodal feature-length behavior are different code paths." + }, + { + "left": "issue:43550", + "right": "issue:45083", + "accept": false, + "reason": "torch.compile/SDPA failure is unrelated to the feature-length helper issue." + }, + { + "left": "issue:45198", + "right": "issue:45803", + "accept": false, + "reason": "Wav2Vec2 save/tokenization failure and Fuyu bare-except masking are distinct tokenizer issues." + }, + { + "left": "issue:45372", + "right": "issue:45588", + "accept": false, + "reason": "Both touch Gemma4-related loading/runtime, but one is an import dependency break and the other is a flash-attention attribute crash." + }, + { + "left": "issue:37428", + "right": "issue:44987", + "accept": false, + "reason": "Missing flash-attention utility import is not the same as failing to load a specific model repository." + }, + { + "left": "issue:29942", + "right": "issue:35141", + "accept": false, + "reason": "Flash Attention 2 test failures are unrelated to embedding reinitialization on resize." + }, + { + "left": "issue:43873", + "right": "issue:44493", + "accept": false, + "reason": "Quantized offloading behavior and unexpected position-id keys are different defects." + }, + { + "left": "issue:45459", + "right": "issue:45712", + "accept": false, + "reason": "Masked tokenizer decode errors and leftover dummy classes are unrelated maintenance issues." + }, + { + "left": "issue:32090", + "right": "issue:33357", + "accept": false, + "reason": "Trainer NoneType broadcast error is unrelated to a MacOS bus error on a CLIP model." + }, + { + "left": "issue:45245", + "right": "issue:45563", + "accept": false, + "reason": "Category-cardinality runtime error is unrelated to a stale warning in paged generate()." + }, + { + "left": "issue:43412", + "right": "issue:45588", + "accept": false, + "reason": "Executorch export failures and flash-attention sink-less crash are different concrete problems." + }, + { + "left": "issue:43873", + "right": "issue:45229", + "accept": false, + "reason": "Quantization/offloading behavior is not the same as multi-GPU OOM on Gemma4." + }, + { + "left": "issue:41720", + "right": "issue:45229", + "accept": false, + "reason": "A800 cudaErrorAssert with auto device mapping is not the same as a multi-GPU OOM." + }, + { + "left": "issue:43278", + "right": "issue:43329", + "accept": false, + "reason": "Embedding dtype drift and undefined video-branch helpers are unrelated model issues." + }, + { + "left": "issue:43976", + "right": "issue:45440", + "accept": false, + "reason": "Python version compatibility and DeepSeek implementation divergence are different failures." + }, + { + "left": "issue:10000", + "right": "issue:15354", + "accept": false, + "reason": "German DistilBertModel issue and torch.jit GeneratorExp export failure are unrelated." + }, + { + "left": "issue:45372", + "right": "issue:45406", + "accept": false, + "reason": "Both involve Gemma4Processor, but the reported failures point to different root causes and code paths." + }, + { + "left": "issue:43065", + "right": "issue:43994", + "accept": false, + "reason": "Dummy Conv2d in Sam3PixelDecoder is unrelated to SigLIP2 producing nonsensical outputs." + }, + { + "left": "issue:42490", + "right": "issue:43575", + "accept": false, + "reason": "Trailing-slash import warning and tensor-parallel OOM are unrelated." + }, + { + "left": "issue:28218", + "right": "issue:29127", + "accept": false, + "reason": "Tokenizer added-token spacing and LayoutLMv3 box-validation messaging are different issues." + }, + { + "left": "issue:43065", + "right": "issue:45412", + "accept": false, + "reason": "Sam3PixelDecoder dummy layer issue is unrelated to RT-DETR memory not releasing." + }, + { + "left": "issue:45372", + "right": "issue:45459", + "accept": false, + "reason": "Gemma4 processor import failure and protobuf-error masking in tokenizers are not the same bug." + }, + { + "left": "issue:44857", + "right": "issue:45588", + "accept": false, + "reason": "AMP/CUDA loss crash and flash-attention sinkless-model crash are different runtime failures." + }, + { + "left": "issue:43653", + "right": "issue:44743", + "accept": false, + "reason": "BigBird special-token registration and Qwen3 recurrent-state reset are unrelated." + }, + { + "left": "issue:44448", + "right": "issue:44466", + "accept": false, + "reason": "Version output differences and lm_head serialization inconsistency are distinct model-state bugs." + }, + { + "left": "issue:43278", + "right": "issue:43381", + "accept": false, + "reason": "Training/eval dtype change and gradient-checkpointing-in-eval are unrelated." + }, + { + "left": "issue:45588", + "right": "issue:45823", + "accept": false, + "reason": "Flash-attention s_aux=None crash and Gemma4 PLE device mismatch are separate Gemma4 runtime bugs." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 26, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:25251", + "issue:29127", + "issue:29942", + "issue:30064", + "issue:32090", + "issue:33290", + "issue:33357", + "issue:34567", + "issue:36010", + "issue:39692", + "issue:42490", + "issue:42915", + "issue:42994", + "issue:43066", + "issue:43257", + "issue:43278", + "issue:43381", + "issue:43421", + "issue:43526", + "issue:43653", + "issue:43701", + "issue:43756", + "issue:43761", + "issue:43824", + "issue:43873", + "issue:43994", + "issue:44186", + "issue:44206", + "issue:44265", + "issue:44279", + "issue:44361", + "issue:44442", + "issue:44451", + "issue:44462", + "issue:44479", + "issue:44610", + "issue:44871", + "issue:44977", + "issue:45103", + "issue:45198", + "issue:45372", + "issue:45412", + "issue:45440", + "issue:45491", + "issue:45520", + "issue:45593", + "issue:45834", + "issue:45859" + ], + "soft_pairs": [ + "issue:44206|issue:44479", + "issue:43066|issue:44462", + "issue:43994|issue:45412", + "issue:25251|issue:32090", + "issue:44186|issue:45198", + "issue:44451|issue:45520", + "issue:42915|issue:42994", + "issue:44361|issue:45198", + "issue:44871|issue:44977", + "issue:44265|issue:45198", + "issue:43756|issue:45440", + "issue:43421|issue:45491", + "issue:29942|issue:30064", + "issue:43257|issue:43278", + "issue:33357|issue:39692", + "issue:44610|issue:45412", + "issue:29127|issue:34567", + "issue:42490|issue:43873", + "issue:32090|issue:33290", + "issue:45103|issue:45372", + "issue:45440|issue:45593", + "issue:43824|issue:44279", + "issue:43653|issue:44442", + "issue:43761|issue:45859", + "issue:43526|issue:45834", + "issue:43381|issue:43701", + "issue:43526|issue:44186", + "issue:29942|issue:36010", + "issue:29127|issue:33357" + ], + "prepared_review_unit_hash": "ee0a4c73a5d587a76ac5f59388e886383b5527958b8abbd48ef3f349c652a9d6", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 29, + "serialized_chars": 36306, + "estimated_input_tokens": 9077, + "estimated_eval_tokens": 18410 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is a broad set of mostly unrelated Transformers issues that only look superficially similar by text embedding. None of the proposed soft edges are true duplicates, so all should be rejected. There is no meaningful canonical issue for the cluster.", + "confidence": 0.92, + "canonical_issue_reason": "No canonical issue: the items span different subsystems and failure modes (tokenization, quantization, training, vision, docs, export, offload), so there is no single underlying bug to represent the cluster.", + "canonical_pr_reason": null, + "best_issue_reason": "No global best issue: the soft matches are all broad similarity false positives rather than one coherent bug family, so no issue is a strong dedupe anchor.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44206", + "right": "issue:44479", + "accept": false, + "reason": "Different regressions in different model paths: LasrFeatureExtractor center-arg handling vs Qwen video-input processing." + }, + { + "left": "issue:43066", + "right": "issue:44462", + "accept": false, + "reason": "Both involve tokenizers, but one is a decoder-type mismatch in v5 while the other is AutoTokenizer ignoring tokenizer.json; not the same bug." + }, + { + "left": "issue:43994", + "right": "issue:45412", + "accept": false, + "reason": "Unrelated failures: SigLIP2 inference corruption vs RT-DETR model memory not being released." + }, + { + "left": "issue:25251", + "right": "issue:32090", + "accept": false, + "reason": "Top-k pipeline output nesting and trainer _gpu_broadcast_one NoneType errors are unrelated code paths." + }, + { + "left": "issue:44186", + "right": "issue:45198", + "accept": false, + "reason": "Both are tokenizer-related, but the concrete failures differ: LayoutLMv2 NER/padding crash vs Wav2Vec2 save_pretrained/tokenization failure." + }, + { + "left": "issue:44451", + "right": "issue:45520", + "accept": false, + "reason": "Different root causes: model loading failure for ScandiBERT vs Python 3.13 flash_attn import-key issue." + }, + { + "left": "issue:42915", + "right": "issue:42994", + "accept": false, + "reason": "Both touch quantization, but one is Qwen3MoE with FineGrainedFP8Config and the other is quantized model saving; not the same defect." + }, + { + "left": "issue:44361", + "right": "issue:45198", + "accept": false, + "reason": "Different tokenizer/model failures affecting different tasks and APIs; no shared underlying bug is evident." + }, + { + "left": "issue:44871", + "right": "issue:44977", + "accept": false, + "reason": "Eos-token config mismatch is not the same issue as Qwen3.5 flash-attention generation failure." + }, + { + "left": "issue:44265", + "right": "issue:45198", + "accept": false, + "reason": "torch.export failure with torch_compilable_check is unrelated to Wav2Vec2 tokenization/save_pretrained problems." + }, + { + "left": "issue:43756", + "right": "issue:45440", + "accept": false, + "reason": "Both concern model implementation divergence, but they affect different architectures and different mechanisms; not mergeable as one fix." + }, + { + "left": "issue:43421", + "right": "issue:45491", + "accept": false, + "reason": "Runtime special-token post-processing is unrelated to Gemma3 NaN embeddings from sliding-window attention." + }, + { + "left": "issue:29942", + "right": "issue:30064", + "accept": false, + "reason": "Flash Attention 2 test failures and void segmentation map image-processing failures are separate bugs." + }, + { + "left": "issue:43257", + "right": "issue:43278", + "accept": false, + "reason": "Different issues: MoE weight conversion with accelerate+deepspeed vs embedding dtype changing between train/eval." + }, + { + "left": "issue:33357", + "right": "issue:39692", + "accept": false, + "reason": "Both mention CLIP/SigLIP-family examples, but one is a Mac bus error on a community model and the other is a documentation example mismatch/quantization issue." + }, + { + "left": "issue:44610", + "right": "issue:45412", + "accept": false, + "reason": "Processor output size mismatch and RT-DETR memory leak are unrelated." + }, + { + "left": "issue:29127", + "right": "issue:34567", + "accept": false, + "reason": "LayoutLMv3 error-message clarity and TrainerState token-count tracking are different issues." + }, + { + "left": "issue:42490", + "right": "issue:43873", + "accept": false, + "reason": "Local-path import warning and quantization offloading behavior are not the same bug." + }, + { + "left": "issue:32090", + "right": "issue:33290", + "accept": false, + "reason": "Trainer NoneType broadcast error and DeepSpeed Adafactor OOM are different failure modes in different layers." + }, + { + "left": "issue:45103", + "right": "issue:45372", + "accept": false, + "reason": "auto_docstring annotation handling and Gemma 4 processor loading/import error are unrelated." + }, + { + "left": "issue:45440", + "right": "issue:45593", + "accept": false, + "reason": "DeepseekV3MoE implementation drift is not the same as D-FINE auxiliary-loss logic when denoising is off." + }, + { + "left": "issue:43824", + "right": "issue:44279", + "accept": false, + "reason": "Qwen2.5-VL import failure is not the same as a generic dependency issue in transformers." + }, + { + "left": "issue:43653", + "right": "issue:44442", + "accept": false, + "reason": "BigBird mask-token registration/empty decode and AutoTokenizer failing to load FastSpeech2ConformerTokenizer are different tokenizer bugs." + }, + { + "left": "issue:43761", + "right": "issue:45859", + "accept": false, + "reason": "CLIPVisionModel hidden_states regression and Qwen3_5Moe CPU-offload incompatibility are unrelated model-forward issues." + }, + { + "left": "issue:43526", + "right": "issue:45834", + "accept": false, + "reason": "BeitImageProcessorFast reduce_labels behavior and Kosmos2.5 OCR index errors are different image-processing problems." + }, + { + "left": "issue:43381", + "right": "issue:43701", + "accept": false, + "reason": "Gradient checkpointing in eval mode and resume_from_checkpoint key mismatch are not the same defect." + }, + { + "left": "issue:43526", + "right": "issue:44186", + "accept": false, + "reason": "ImageProcessorFast label reduction and LayoutLMv2 tokenizer crashes are different components and failure modes." + }, + { + "left": "issue:29942", + "right": "issue:36010", + "accept": false, + "reason": "Flash Attention 2 test failures are unrelated to an ImportError for GenerationMixin." + }, + { + "left": "issue:29127", + "right": "issue:33357", + "accept": false, + "reason": "LayoutLMv3 error-message clarity and a MacOS CLIP bus error are unrelated." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 27, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:29127", + "issue:33290", + "issue:34567", + "issue:35141", + "issue:42490", + "issue:42907", + "issue:42915", + "issue:43232", + "issue:43262", + "issue:43316", + "issue:43334", + "issue:43381", + "issue:43388", + "issue:43441", + "issue:43525", + "issue:43526", + "issue:43531", + "issue:43575", + "issue:43646", + "issue:43701", + "issue:43723", + "issue:43761", + "issue:43931", + "issue:44062", + "issue:44077", + "issue:44079", + "issue:44188", + "issue:44206", + "issue:44387", + "issue:44442", + "issue:44464", + "issue:44479", + "issue:44743", + "issue:44811", + "issue:44857", + "issue:44898", + "issue:44977", + "issue:45003", + "issue:45072", + "issue:45245", + "issue:45290", + "issue:45440", + "issue:45479", + "issue:45491", + "issue:45588", + "issue:45698", + "issue:45803", + "issue:45907" + ], + "soft_pairs": [ + "issue:43334|issue:43531", + "issue:44977|issue:45245", + "issue:29127|issue:35141", + "issue:44464|issue:45491", + "issue:42915|issue:43931", + "issue:44442|issue:45072", + "issue:43262|issue:44206", + "issue:43526|issue:44857", + "issue:42490|issue:43381", + "issue:44062|issue:45907", + "issue:43232|issue:43388", + "issue:43441|issue:45588", + "issue:44743|issue:44811", + "issue:43761|issue:44079", + "issue:43646|issue:44077", + "issue:43575|issue:43701", + "issue:45003|issue:45440", + "issue:43575|issue:44387", + "issue:43723|issue:45479", + "issue:44442|issue:45803", + "issue:43761|issue:44188", + "issue:44479|issue:45290", + "issue:43316|issue:43525", + "issue:42907|issue:45698", + "issue:44442|issue:44898", + "issue:42490|issue:44743", + "issue:33290|issue:34567" + ], + "prepared_review_unit_hash": "4dfbf2e175a9b36c154645749ed5b3373096da7a26b298968905913070f7320e", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 27, + "serialized_chars": 35884, + "estimated_input_tokens": 8971, + "estimated_eval_tokens": 18198 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The items are mostly unrelated one-off bugs/regressions across different subsystems; the soft pairs share superficial vocabulary but not the same concrete failure mode. I would reject all soft edges.", + "confidence": 0.96, + "canonical_issue_reason": "issue:45698 is the broadest, most general loading/saving bug in the set and is still open; the rest are mostly model-specific or subsystem-specific regressions that do not form a clear duplicate cluster.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45698 is the best representative if one issue must anchor this cluster: it concerns core from_pretrained/save_pretrained behavior and has broader reuse potential than the other narrowly scoped bugs.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43334", + "right": "issue:43531", + "accept": false, + "reason": "Different Qwen issues: missing `pad_token_id` on Qwen3-VL vs `sliding_window` behavior on Qwen3-MoE; not the same bug." + }, + { + "left": "issue:44977", + "right": "issue:45245", + "accept": false, + "reason": "Flash-attention generation regression vs a generic category-count runtime limit; unrelated failure modes." + }, + { + "left": "issue:29127", + "right": "issue:35141", + "accept": false, + "reason": "LayoutLMv3 box-clarity error message vs token-embedding reinitialization after resize; different code paths and symptoms." + }, + { + "left": "issue:44464", + "right": "issue:45491", + "accept": false, + "reason": "Compiled chunked generation inconsistency vs Gemma3 NaN embeddings from sliding-window/all-padding windows; not the same defect." + }, + { + "left": "issue:42915", + "right": "issue:43931", + "accept": false, + "reason": "Qwen3Moe FineGrainedFP8Config failure vs Qwen3-VL weight-shape mismatch; separate model-loading problems." + }, + { + "left": "issue:44442", + "right": "issue:45072", + "accept": false, + "reason": "Tokenizer loading failure for FastSpeech2ConformerTokenizer vs dtype mismatch in SwitchTransformers/TimmWrapperModel inference." + }, + { + "left": "issue:43262", + "right": "issue:44206", + "accept": false, + "reason": "Audio chat-template sampling-rate default issue vs unsupported `center` argument in a different feature extractor." + }, + { + "left": "issue:43526", + "right": "issue:44857", + "accept": false, + "reason": "BEiTImageProcessorFast label reduction bug vs LwDetrImageLoss AMP/CUDA crash; unrelated subsystems." + }, + { + "left": "issue:42490", + "right": "issue:43381", + "accept": false, + "reason": "Trailing-slash local model import warning vs gradient checkpointing being disallowed in eval mode; not the same bug." + }, + { + "left": "issue:44062", + "right": "issue:45907", + "accept": false, + "reason": "`AddedToken` keyword collision vs `list_repo_files` not forwarding `revision`; different APIs and failures." + }, + { + "left": "issue:43232", + "right": "issue:43388", + "accept": false, + "reason": "Generation kwargs update after `sync_gpus` vs metric gathering dropping tuple labels; separate bugs." + }, + { + "left": "issue:43441", + "right": "issue:45588", + "accept": false, + "reason": "Both involve flash attention, but one is a Ministral-3 model regression and the other is an `s_aux=None` crash in a different integration path." + }, + { + "left": "issue:44743", + "right": "issue:44811", + "accept": false, + "reason": "Recurrent state reset in Qwen3.5 MoE vs Whisper `batch_decode` ignoring `skip_special_tokens`; no shared underlying bug." + }, + { + "left": "issue:43761", + "right": "issue:44079", + "accept": false, + "reason": "CLIPVisionModel not returning hidden states vs ModelOutput key assignment when a key was previously `None`; different abstractions." + }, + { + "left": "issue:43646", + "right": "issue:44077", + "accept": false, + "reason": "Transformers 5.0 custom model init breakage vs `patchtsmixer` `post_init` contract issue; not the same problem." + }, + { + "left": "issue:43575", + "right": "issue:43701", + "accept": false, + "reason": "TP loading OOM for a Qwen2 model vs checkpoint resume key mismatch; unrelated." + }, + { + "left": "issue:45003", + "right": "issue:45440", + "accept": false, + "reason": "Unsafe `sys.modules` access in `modeling_utils` vs DeepSeekV3MoE implementation divergence; different code paths." + }, + { + "left": "issue:43575", + "right": "issue:44387", + "accept": false, + "reason": "Both mention OOM, but one is TP model loading and the other is int4 quantization reserved-memory growth; not mergeable as one bug." + }, + { + "left": "issue:43723", + "right": "issue:45479", + "accept": false, + "reason": "AutoTokenizer loading issue in v5 vs degenerate zero-loss sequence classification; unrelated." + }, + { + "left": "issue:44442", + "right": "issue:45803", + "accept": false, + "reason": "Tokenizer loading failure vs bare `except` in Fuyu tensor conversion; no shared underlying defect." + }, + { + "left": "issue:43761", + "right": "issue:44188", + "accept": false, + "reason": "Hidden-states regression in CLIPVisionModel vs attention-kernel divergence under `torch.compile`; distinct issues." + }, + { + "left": "issue:44479", + "right": "issue:45290", + "accept": false, + "reason": "Video-input regression for Qwen-family models vs chat-template tokenization crash on tool-call assistant messages; different layers." + }, + { + "left": "issue:43316", + "right": "issue:43525", + "accept": false, + "reason": "Gemma3TextConfig API discrepancy vs Llama4Config missing `pad_token_id`; similar theme, but not the same bug." + }, + { + "left": "issue:42907", + "right": "issue:45698", + "accept": false, + "reason": "Both touch model save/load, but one is dequantized Ministral save failure and the other is wrong custom module loading after save/load; too different to merge." + }, + { + "left": "issue:44442", + "right": "issue:44898", + "accept": false, + "reason": "Tokenizer loading failure vs Perceiver non-default-resolution image classification failure; unrelated." + }, + { + "left": "issue:42490", + "right": "issue:44743", + "accept": false, + "reason": "Local path import warning vs Qwen3.5 recurrent-state reset bug; no common failure mode." + }, + { + "left": "issue:33290", + "right": "issue:34567", + "accept": false, + "reason": "DeepSpeed Adafactor OOM vs `num_input_tokens_seen` not updating; entirely separate training issues." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 28, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:25251", + "issue:28282", + "issue:33290", + "issue:34689", + "issue:35141", + "issue:36010", + "issue:36331", + "issue:39692", + "issue:41628", + "issue:42175", + "issue:43065", + "issue:43295", + "issue:43316", + "issue:43381", + "issue:43388", + "issue:43404", + "issue:43441", + "issue:43493", + "issue:43540", + "issue:43611", + "issue:43643", + "issue:43673", + "issue:43688", + "issue:43720", + "issue:43749", + "issue:43825", + "issue:43827", + "issue:43873", + "issue:43931", + "issue:44077", + "issue:44162", + "issue:44190", + "issue:44442", + "issue:44479", + "issue:44496", + "issue:44514", + "issue:44617", + "issue:44743", + "issue:44792", + "issue:44855", + "issue:44871", + "issue:44877", + "issue:44913", + "issue:45137", + "issue:45216", + "issue:45372", + "issue:45491", + "issue:45759" + ], + "soft_pairs": [ + "issue:34689|issue:43611", + "issue:39692|issue:45491", + "issue:33290|issue:35141", + "issue:43931|issue:45216", + "issue:33290|issue:34689", + "issue:43381|issue:43873", + "issue:43493|issue:43827", + "issue:44077|issue:44479", + "issue:43316|issue:45759", + "issue:43720|issue:44442", + "issue:41628|issue:45372", + "issue:43643|issue:44913", + "issue:43749|issue:43931", + "issue:34689|issue:36010", + "issue:43873|issue:44871", + "issue:44190|issue:44496", + "issue:33290|issue:36331", + "issue:44855|issue:45137", + "issue:43316|issue:44877", + "issue:43540|issue:44514", + "issue:42175|issue:43316", + "issue:43673|issue:44743", + "issue:43065|issue:44617", + "issue:43404|issue:43441", + "issue:44162|issue:44792", + "issue:43295|issue:43825", + "issue:25251|issue:28282", + "issue:43388|issue:43688" + ], + "prepared_review_unit_hash": "1a7b2f9b063d6aaf2222e6d3d8f35b2cb02f0997eaa0d4afcecb595a62dce6fa", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 28, + "serialized_chars": 36336, + "estimated_input_tokens": 9084, + "estimated_eval_tokens": 18424 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a false-positive semantic neighborhood: the issues span unrelated regressions, model families, and API surfaces. I found no pair that looks like the same underlying bug/change, so all soft edges should be rejected.", + "confidence": 0.95, + "canonical_issue_reason": "No true duplicate center emerges; if one issue must be chosen, issue 34689 is a clear, well-described regression report and a reasonable representative of the cluster\u2019s broad 'model loading breakage' theme.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 34689 is the most concrete and broadly representative bug report in an otherwise heterogeneous set, but it is not a duplicate of the others.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:34689", + "right": "issue:43611", + "accept": false, + "reason": "Both involve loading regressions, but one is a specific Llama 3.2 Vision breakage in 4.46.2 and the other is a v5 base_model_prefix loading change; different code paths and failure modes." + }, + { + "left": "issue:39692", + "right": "issue:45491", + "accept": false, + "reason": "Different models and different bugs: SigLIP2 docs/example errors versus Gemma3 NaN embeddings during mixed-length batching." + }, + { + "left": "issue:33290", + "right": "issue:35141", + "accept": false, + "reason": "OOM with Adafactor + DeepSpeed is unrelated to token embedding resize reinitialization in post_init." + }, + { + "left": "issue:43931", + "right": "issue:45216", + "accept": false, + "reason": "One is a weight-shape mismatch during loading, the other is an incorrect save_pretrained checkpoint for Qwen3.5; not the same underlying bug." + }, + { + "left": "issue:33290", + "right": "issue:34689", + "accept": false, + "reason": "Different subsystems and symptoms: optimizer/deepspeed memory blowup versus model-loading regression for a specific vision model." + }, + { + "left": "issue:43381", + "right": "issue:43873", + "accept": false, + "reason": "Gradient checkpointing in eval mode is unrelated to quantization/offloading behavior." + }, + { + "left": "issue:43493", + "right": "issue:43827", + "accept": false, + "reason": "SigLIP2 implementation discrepancy and stale pipeline documentation are different issues." + }, + { + "left": "issue:44077", + "right": "issue:44479", + "accept": false, + "reason": "PatchTSMixer post_init validation and Qwen video-input regression are unrelated." + }, + { + "left": "issue:43316", + "right": "issue:45759", + "accept": false, + "reason": "Both touch config APIs, but one is a Gemma3TextConfig inconsistency and the other is AutoModelForCausalLM unwrapping composite multimodal text_config; different fixes." + }, + { + "left": "issue:43720", + "right": "issue:44442", + "accept": false, + "reason": "BitNet accelerate loading with packed weights is unrelated to AutoTokenizer failing for a custom tokenizer." + }, + { + "left": "issue:41628", + "right": "issue:45372", + "accept": false, + "reason": "Both are import errors, but they concern different missing symbols and different dependency chains." + }, + { + "left": "issue:43643", + "right": "issue:44913", + "accept": false, + "reason": "Trust-remote-code config field loss is not the same as GPTNeoX rotary_pct resetting on reload." + }, + { + "left": "issue:43749", + "right": "issue:43931", + "accept": false, + "reason": "FSDP CPU RAM efficient loading failure is not the same as a specific model weight-shape mismatch." + }, + { + "left": "issue:34689", + "right": "issue:36010", + "accept": false, + "reason": "Different loading errors with different root causes; one is a version regression for Llama 3.2 Vision, the other is a missing GenerationMixin import path." + }, + { + "left": "issue:43873", + "right": "issue:44871", + "accept": false, + "reason": "Quantization/offloading behavior and Gemma-3 eos_token_id inconsistency are unrelated." + }, + { + "left": "issue:44190", + "right": "issue:44496", + "accept": false, + "reason": "Dataset loading in a training script is unrelated to unrecognized model_type in config.json." + }, + { + "left": "issue:33290", + "right": "issue:36331", + "accept": false, + "reason": "OOM in DeepSpeed/Adafactor is unrelated to CustomTrainer.compute_loss signature mismatch." + }, + { + "left": "issue:44855", + "right": "issue:45137", + "accept": false, + "reason": "Python 3.13 IndentationError on import is unrelated to a DeepSpeed ZeRO3 deque underflow." + }, + { + "left": "issue:43316", + "right": "issue:44877", + "accept": false, + "reason": "Both involve configs, but the specific discrepancies and affected models differ; not the same bug." + }, + { + "left": "issue:43540", + "right": "issue:44514", + "accept": false, + "reason": "Both are multimodal processor bugs, but one is Qwen3OmniMoe video processing and the other is batched apply_chat_template padding behavior in Qwen2_5_VL." + }, + { + "left": "issue:42175", + "right": "issue:43316", + "accept": false, + "reason": "Packaging/backend dependency omission is unrelated to a Gemma3TextConfig API discrepancy." + }, + { + "left": "issue:43673", + "right": "issue:44743", + "accept": false, + "reason": "Both mention cache, but the first is missing GenerationMixin cache in chunked_prefill while the second is recurrent state reset in modular qwen3_5; different code paths." + }, + { + "left": "issue:43065", + "right": "issue:44617", + "accept": false, + "reason": "A dummy Conv2d implementation issue is not the same as a CUDA OOM in Sam3Video." + }, + { + "left": "issue:43404", + "right": "issue:43441", + "accept": false, + "reason": "Mistral3 lm_head tying and Ministral-3 FlashAttention failure are distinct bugs." + }, + { + "left": "issue:44162", + "right": "issue:44792", + "accept": false, + "reason": "ESM2 breakage affecting workflows and a janus image-generation test failure are unrelated." + }, + { + "left": "issue:43295", + "right": "issue:43825", + "accept": false, + "reason": "Processor.tokenizer/image passing regression is different from a pipeline error message about translation support." + }, + { + "left": "issue:25251", + "right": "issue:28282", + "accept": false, + "reason": "Pipeline top_k nesting behavior is unrelated to a missing PyTorch dependency ImportError." + }, + { + "left": "issue:43388", + "right": "issue:43688", + "accept": false, + "reason": "Label truncation in gather_for_metrics is unrelated to auxiliary-loss normalization in OLMoE/GPT Oss." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 29, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:16998", + "issue:22355", + "issue:32090", + "issue:35141", + "issue:41950", + "issue:42617", + "issue:43065", + "issue:43295", + "issue:43452", + "issue:43475", + "issue:43504", + "issue:43550", + "issue:43575", + "issue:43606", + "issue:43646", + "issue:43701", + "issue:43742", + "issue:43761", + "issue:43827", + "issue:43906", + "issue:43931", + "issue:43994", + "issue:44060", + "issue:44077", + "issue:44190", + "issue:44242", + "issue:44464", + "issue:44479", + "issue:44496", + "issue:44521", + "issue:44534", + "issue:44610", + "issue:44617", + "issue:44734", + "issue:44792", + "issue:44871", + "issue:44913", + "issue:44964", + "issue:44977", + "issue:45127", + "issue:45200", + "issue:45290", + "issue:45381", + "issue:45406", + "issue:45563", + "issue:45584", + "issue:45698" + ], + "soft_pairs": [ + "issue:41950|issue:43475", + "issue:44077|issue:44534", + "issue:16998|issue:22355", + "issue:43550|issue:43701", + "issue:44871|issue:45584", + "issue:43761|issue:44479", + "issue:43606|issue:43701", + "issue:45127|issue:45200", + "issue:43295|issue:44871", + "issue:43452|issue:44496", + "issue:42617|issue:43575", + "issue:43575|issue:43742", + "issue:44734|issue:45406", + "issue:43646|issue:44913", + "issue:43504|issue:45698", + "issue:44617|issue:45584", + "issue:43761|issue:44242", + "issue:43931|issue:44060", + "issue:45563|issue:45584", + "issue:32090|issue:35141", + "issue:44464|issue:44871", + "issue:44479|issue:44734", + "issue:43065|issue:44610", + "issue:44792|issue:44977", + "issue:44521|issue:45381", + "issue:43906|issue:45290", + "issue:43827|issue:43994", + "issue:44464|issue:45563", + "issue:44534|issue:44913", + "issue:44190|issue:44964" + ], + "prepared_review_unit_hash": "2e4e383b1b7301f3b6a9c5a6cbf4247fa6bce2ad146eaaefa5fca5a9948f2d87", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 30, + "serialized_chars": 36073, + "estimated_input_tokens": 9019, + "estimated_eval_tokens": 18294 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The set is heterogeneous: most soft pairs share only broad subsystem or wording similarity, not the same underlying bug. No pull requests are present. Issue 44521 is the best anchor only because it is open, active, and well-scoped.", + "confidence": 0.42, + "canonical_issue_reason": "Issue 44521 is the strongest canonical anchor in this set: it is open, has the most inbound references, and describes a concrete reproducible multimodal chat-template regression.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 44521 is the most triage-worthy representative issue because it is active, well-described, and likely to attract follow-up reports.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:41950", + "right": "issue:43475", + "accept": false, + "reason": "Different bugs: pipeline processor lookup vs missing model output attribute in SAM3 video." + }, + { + "left": "issue:44077", + "right": "issue:44534", + "accept": false, + "reason": "One is a post_init allowance policy change; the other is a v5 buffer initialization regression." + }, + { + "left": "issue:16998", + "right": "issue:22355", + "accept": false, + "reason": "Unrelated topics: DeBERTa model_max_length question vs missing transformers.onnx module." + }, + { + "left": "issue:43550", + "right": "issue:43701", + "accept": false, + "reason": "SDPA/torch.compile failure is unrelated to resume_from_checkpoint key mismatch." + }, + { + "left": "issue:44871", + "right": "issue:45584", + "accept": false, + "reason": "Different code paths: eos_token_id config mismatch vs Whisper empty-transcription generation failure." + }, + { + "left": "issue:43761", + "right": "issue:44479", + "accept": false, + "reason": "Both are regressions, but on different models and different failures; not the same bug." + }, + { + "left": "issue:43606", + "right": "issue:43701", + "accept": false, + "reason": "CPU offload device mismatch is unrelated to checkpoint key mismatch." + }, + { + "left": "issue:45127", + "right": "issue:45200", + "accept": false, + "reason": "LoRA merge / tied embeddings issue is different from Gemma 4 mm_token_type_ids defaults." + }, + { + "left": "issue:43295", + "right": "issue:44871", + "accept": false, + "reason": "Processor/tokenizer regression is unrelated to eos_token_id configuration drift." + }, + { + "left": "issue:43452", + "right": "issue:44496", + "accept": false, + "reason": "Both concern loading, but the root causes differ: gguf_file handling vs missing model_type in config." + }, + { + "left": "issue:42617", + "right": "issue:43575", + "accept": false, + "reason": "A script startup failure and an OOM during tensor-parallel loading are not the same defect." + }, + { + "left": "issue:43575", + "right": "issue:43742", + "accept": false, + "reason": "OOM during model load does not match a key error for MobileLLM loading." + }, + { + "left": "issue:44734", + "right": "issue:45406", + "accept": false, + "reason": "Both hit transformers serve, but one is KV-cache indexing and the other is a missing processor attribute." + }, + { + "left": "issue:43646", + "right": "issue:44913", + "accept": false, + "reason": "Custom model initialization breakage is not the same as GPTNeoXConfig reload reverting rotary_pct." + }, + { + "left": "issue:43504", + "right": "issue:45698", + "accept": false, + "reason": "Legacy field handling in BEiT preset loading is unrelated to wrong custom-module selection after save_pretrained." + }, + { + "left": "issue:44617", + "right": "issue:45584", + "accept": false, + "reason": "CUDA OOM in Sam3Video is unrelated to Whisper empty transcription behavior." + }, + { + "left": "issue:43761", + "right": "issue:44242", + "accept": false, + "reason": "Hidden_states missing in CLIPVisionModel is different from missing load-balancing loss when router logits are off." + }, + { + "left": "issue:43931", + "right": "issue:44060", + "accept": false, + "reason": "Shape mismatch on Qwen3-VL load is not the same as tied-weights warning in Qwen3-Next." + }, + { + "left": "issue:45563", + "right": "issue:45584", + "accept": false, + "reason": "A stale warning in paged generate is unrelated to a generation crash on empty Whisper output." + }, + { + "left": "issue:32090", + "right": "issue:35141", + "accept": false, + "reason": "Trainer broadcast NoneType error is unrelated to embedding reinitialization after resize." + }, + { + "left": "issue:44464", + "right": "issue:44871", + "accept": false, + "reason": "Chunked generation inconsistency with compiled forward is unrelated to eos_token_id config inconsistency." + }, + { + "left": "issue:44479", + "right": "issue:44734", + "accept": false, + "reason": "Both are regressions, but one concerns video input processing and the other server-side KV-cache indexing." + }, + { + "left": "issue:43065", + "right": "issue:44610", + "accept": false, + "reason": "Dummy Conv2d in Sam3PixelDecoder is different from OmDet-Turbo processor/model input-size mismatch." + }, + { + "left": "issue:44792", + "right": "issue:44977", + "accept": false, + "reason": "Failed Janus image-generation test is unrelated to Qwen3.5 flash-attention generation." + }, + { + "left": "issue:44521", + "right": "issue:45381", + "accept": false, + "reason": "Both are multimodal, but assistant mask generation and video vision_position_ids are different bugs." + }, + { + "left": "issue:43906", + "right": "issue:45290", + "accept": false, + "reason": "An isolated reproduction of an older issue is not enough to match the tool-call/no-content chat-template crash." + }, + { + "left": "issue:43827", + "right": "issue:43994", + "accept": false, + "reason": "Documentation references to removed pipelines do not match a model-output correctness problem." + }, + { + "left": "issue:44464", + "right": "issue:45563", + "accept": false, + "reason": "Generation inconsistency and a stale warning are different severity and different failure modes." + }, + { + "left": "issue:44534", + "right": "issue:44913", + "accept": false, + "reason": "Non-persistent buffer junk in v5 is unrelated to GPTNeoX config reload defaults." + }, + { + "left": "issue:44190", + "right": "issue:44964", + "accept": false, + "reason": "Local dataset loading in an example script is unrelated to loading a multimodal model preset." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 30, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:35141", + "issue:36246", + "issue:38175", + "issue:38617", + "issue:39692", + "issue:41628", + "issue:41720", + "issue:41950", + "issue:42490", + "issue:43065", + "issue:43295", + "issue:43335", + "issue:43388", + "issue:43479", + "issue:43540", + "issue:43550", + "issue:43644", + "issue:43650", + "issue:43976", + "issue:44016", + "issue:44077", + "issue:44079", + "issue:44112", + "issue:44186", + "issue:44190", + "issue:44220", + "issue:44462", + "issue:44496", + "issue:44734", + "issue:44737", + "issue:44792", + "issue:44855", + "issue:44913", + "issue:44933", + "issue:44938", + "issue:45127", + "issue:45245", + "issue:45276", + "issue:45357", + "issue:45362", + "issue:45375", + "issue:45542", + "issue:45584", + "issue:45588", + "issue:45600", + "issue:45803", + "issue:45907" + ], + "soft_pairs": [ + "issue:44933|issue:45362", + "issue:41720|issue:45245", + "issue:44792|issue:45600", + "issue:44855|issue:45588", + "issue:44737|issue:45803", + "issue:45127|issue:45357", + "issue:43388|issue:44079", + "issue:35141|issue:42490", + "issue:44462|issue:45907", + "issue:43335|issue:44855", + "issue:44077|issue:44079", + "issue:44938|issue:45542", + "issue:43550|issue:44190", + "issue:43976|issue:44855", + "issue:43644|issue:44913", + "issue:43540|issue:44734", + "issue:44112|issue:45600", + "issue:44496|issue:45803", + "issue:43335|issue:44190", + "issue:43650|issue:44220", + "issue:45276|issue:45584", + "issue:39692|issue:44190", + "issue:45245|issue:45584", + "issue:41950|issue:44016", + "issue:43479|issue:45375", + "issue:43295|issue:45276", + "issue:36246|issue:41628", + "issue:36246|issue:38617", + "issue:44186|issue:44734", + "issue:38175|issue:43065" + ], + "prepared_review_unit_hash": "a5a112a68f564965a0408cf3f9900ba9ba3db224c9f535914556750a49544df1", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 30, + "serialized_chars": 36323, + "estimated_input_tokens": 9081, + "estimated_eval_tokens": 18418 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The soft pairs are mostly high-similarity false positives: they overlap in broad subsystem vocabulary (imports, configs, pipelines, CI, model loading) but describe different models, code paths, and failure modes. I would not deduplicate any of them.", + "confidence": 0.97, + "canonical_issue_reason": "If a single issue must be chosen as the representative, issue 45276 is the clearest, most concrete standalone bug in this set: it precisely describes a resize-token-embeddings/output-embeddings mismatch with a reproducible model behavior change.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 45276 is the strongest issue candidate because it is specific, actionable, and tied to a concrete state inconsistency after embedding resizing; it is the most understandable representative among otherwise unrelated reports.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44933", + "right": "issue:45362", + "accept": false, + "reason": "Different problems: an invalid import from image_utils vs a Qwen3.5 chat runtime crash." + }, + { + "left": "issue:41720", + "right": "issue:45245", + "accept": false, + "reason": "Unrelated failures: auto device mapping CUDA assert vs category-count limit error." + }, + { + "left": "issue:44792", + "right": "issue:45600", + "accept": false, + "reason": "Both are CI-ish, but one is a janus image-generation test failure and the other is a stale auto-mapping reference bug." + }, + { + "left": "issue:44855", + "right": "issue:45588", + "accept": false, + "reason": "Completely different code paths: Python 3.13 parsing/IndentationError vs flash-attention aux-state handling." + }, + { + "left": "issue:44737", + "right": "issue:45803", + "accept": false, + "reason": "Different bugs: XLNet device placement in positional encoding vs a bare-except swallowing errors in Fuyu tensor conversion." + }, + { + "left": "issue:45127", + "right": "issue:45357", + "accept": false, + "reason": "Same model family, but distinct issues: LoRA merge with extended vocab vs incorrect visual encoder keys in save_pretrained." + }, + { + "left": "issue:43388", + "right": "issue:44079", + "accept": false, + "reason": "One is metric gathering dropping tuple labels; the other is ModelOutput field assignment with None keys." + }, + { + "left": "issue:35141", + "right": "issue:42490", + "accept": false, + "reason": "Embedding reinitialization after resize is unrelated to the local-path trailing-slash import warning." + }, + { + "left": "issue:44462", + "right": "issue:45907", + "accept": false, + "reason": "Both involve repository/file lookup, but tokenizer.json precedence and list_repo_files revision forwarding are different APIs and failures." + }, + { + "left": "issue:43335", + "right": "issue:44855", + "accept": false, + "reason": "Different models and bugs: SwitchTransformers sparse-layer creation vs DebertaV2 Python 3.13 import parsing." + }, + { + "left": "issue:44077", + "right": "issue:44079", + "accept": false, + "reason": "Unrelated: patchtsmixer post_init policy vs ModelOutput key reassignment behavior." + }, + { + "left": "issue:44938", + "right": "issue:45542", + "accept": false, + "reason": "Python 3.14 import/load failure is not the same as the tensorboard-without-TensorFlow backend error." + }, + { + "left": "issue:43550", + "right": "issue:44190", + "accept": false, + "reason": "Different surfaces entirely: torch.compile/SDPA failure in Bamba vs dataset loading in an image-classification script." + }, + { + "left": "issue:43976", + "right": "issue:44855", + "accept": false, + "reason": "Both are version-compatibility reports, but one targets Python 3.9/3.10 behavior and the other is a Python 3.13 IndentationError." + }, + { + "left": "issue:43644", + "right": "issue:44913", + "accept": false, + "reason": "Non-persistent buffer initialization junk is not the same as GPTNeoX rotary_pct not persisting on reload." + }, + { + "left": "issue:43540", + "right": "issue:44734", + "accept": false, + "reason": "Different runtime paths: Qwen3OmniMoe video-input validation vs transformers serve KV-cache indexing." + }, + { + "left": "issue:44112", + "right": "issue:45600", + "accept": false, + "reason": "A stale GraniteSpeech test is not the same as broken auto_mappings references to removed configs." + }, + { + "left": "issue:44496", + "right": "issue:45803", + "accept": false, + "reason": "Model config discovery/import failure is unrelated to FuyuBatchFeature's broad exception handling." + }, + { + "left": "issue:43335", + "right": "issue:44190", + "accept": false, + "reason": "Different subsystems and symptoms: SwitchTransformers config creation vs local dataset loading in a training script." + }, + { + "left": "issue:43650", + "right": "issue:44220", + "accept": false, + "reason": "The placeholder 'ADD THE DATA' issue is not the same as an audio feature extraction bug." + }, + { + "left": "issue:45276", + "right": "issue:45584", + "accept": false, + "reason": "Embedding-resize/output-embedding inconsistency is unrelated to Whisper empty-transcription generation after token alignment." + }, + { + "left": "issue:39692", + "right": "issue:44190", + "accept": false, + "reason": "Docs example/quantization failures are distinct from a local dataset loading problem." + }, + { + "left": "issue:45245", + "right": "issue:45584", + "accept": false, + "reason": "Different errors and models: category cardinality limit vs Whisper generation on empty transcription." + }, + { + "left": "issue:41950", + "right": "issue:44016", + "accept": false, + "reason": "One is a video-classification pipeline bug; the other is a notebook syntax error." + }, + { + "left": "issue:43479", + "right": "issue:45375", + "accept": false, + "reason": "Both are config issues, but they affect different multimodal configs and different fields/defaults." + }, + { + "left": "issue:43295", + "right": "issue:45276", + "accept": false, + "reason": "Regression in processor.tokenizer/image passing is unrelated to Gemma4 token-embedding resize behavior." + }, + { + "left": "issue:36246", + "right": "issue:41628", + "accept": false, + "reason": "Both are import errors, but they involve different symbols and different model modules." + }, + { + "left": "issue:36246", + "right": "issue:38617", + "accept": false, + "reason": "Different missing imports from different modules; not the same underlying bug." + }, + { + "left": "issue:44186", + "right": "issue:44734", + "accept": false, + "reason": "Tokenizer padding/truncation crash and serve KV-cache indexing crash are separate failures." + }, + { + "left": "issue:38175", + "right": "issue:43065", + "accept": false, + "reason": "Different models and mechanisms: SigLIP2 zero probabilities vs Sam3PixelDecoder dummy Conv2d." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 31, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:42175", + "issue:42490", + "issue:42907", + "issue:43295", + "issue:43299", + "issue:43381", + "issue:43404", + "issue:43441", + "issue:43526", + "issue:43531", + "issue:43643", + "issue:43673", + "issue:43824", + "issue:43854", + "issue:43901", + "issue:43976", + "issue:44016", + "issue:44062", + "issue:44075", + "issue:44112", + "issue:44220", + "issue:44246", + "issue:44279", + "issue:44336", + "issue:44351", + "issue:44368", + "issue:44466", + "issue:44496", + "issue:44561", + "issue:44683", + "issue:44734", + "issue:44913", + "issue:44933", + "issue:44964", + "issue:44987", + "issue:45003", + "issue:45042", + "issue:45216", + "issue:45278", + "issue:45335", + "issue:45357", + "issue:45446", + "issue:45584", + "issue:45593", + "issue:45600", + "issue:45632", + "issue:45759" + ], + "soft_pairs": [ + "issue:44683|issue:45446", + "issue:42175|issue:44279", + "issue:44016|issue:44220", + "issue:44466|issue:44913", + "issue:44734|issue:45632", + "issue:43404|issue:45216", + "issue:44561|issue:45632", + "issue:44246|issue:45003", + "issue:42490|issue:45632", + "issue:43643|issue:44496", + "issue:44964|issue:45600", + "issue:44112|issue:44964", + "issue:42907|issue:43441", + "issue:43854|issue:44964", + "issue:44016|issue:44062", + "issue:43976|issue:44987", + "issue:43295|issue:45335", + "issue:43673|issue:44336", + "issue:43531|issue:45600", + "issue:43824|issue:44933", + "issue:44496|issue:45042", + "issue:43526|issue:43901", + "issue:43299|issue:45216", + "issue:44075|issue:44368", + "issue:45335|issue:45584", + "issue:43976|issue:44279", + "issue:43381|issue:45593", + "issue:43299|issue:45357", + "issue:43901|issue:44496", + "issue:44496|issue:45759", + "issue:44351|issue:45278" + ], + "prepared_review_unit_hash": "0b8816fae2909264dc6cae672aa096cccdca1187a24b84a208a3bc33b3bcc904", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 31, + "serialized_chars": 36350, + "estimated_input_tokens": 9088, + "estimated_eval_tokens": 18432 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "These items are a heterogeneous set of unrelated Transformers issues; the soft pairs mostly only share broad subsystem or regression language, not the same concrete bug. No soft edge should be merged.", + "confidence": 0.97, + "canonical_issue_reason": "No true duplicate cluster emerges from these candidates. If one issue must serve as an umbrella, issue 45278 is the broadest (many import errors after upgrade), but it is still not the same underlying bug as the others.", + "canonical_pr_reason": null, + "best_issue_reason": "issue 45278 is the most general and umbrella-like issue in the set, so it is the least bad anchor for triage, though the set is too diverse to treat as one duplicate cluster.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44683", + "right": "issue:45446", + "accept": false, + "reason": "Different failures: compiled flex_attention on torch>=2.9 vs a wrong PyTorch version check for AuxRequest import." + }, + { + "left": "issue:42175", + "right": "issue:44279", + "accept": false, + "reason": "Packaging dependency problems, but one is missing TensorFlow in extras and the other is a general transformers dependency issue; not the same bug." + }, + { + "left": "issue:44016", + "right": "issue:44220", + "accept": false, + "reason": "Notebook syntax error vs a feature-specific audio feature extraction issue; unrelated code paths." + }, + { + "left": "issue:44466", + "right": "issue:44913", + "accept": false, + "reason": "Both involve config/serialization behavior, but one is tied lm_head.weight saving and the other is rotary_pct not persisting on reload." + }, + { + "left": "issue:44734", + "right": "issue:45632", + "accept": false, + "reason": "Serve/KV-cache tensor indexing crash is unrelated to a trust_remote_code cache-path collision for local models." + }, + { + "left": "issue:43404", + "right": "issue:45216", + "accept": false, + "reason": "Weight tying in Mistral3 conditional generation is unrelated to Qwen3.5 save_pretrained checkpoint corruption." + }, + { + "left": "issue:44561", + "right": "issue:45632", + "accept": false, + "reason": "Removal of a helper breaking trust_remote_code models is a different failure mode from local cache-path collisions." + }, + { + "left": "issue:44246", + "right": "issue:45003", + "accept": false, + "reason": "Import latency during transformers import is not the same as unsafe sys.modules access in modeling_utils." + }, + { + "left": "issue:42490", + "right": "issue:45632", + "accept": false, + "reason": "Local path ending with slash import warning is different from cache path collisions between same leaf directory names." + }, + { + "left": "issue:43643", + "right": "issue:44496", + "accept": false, + "reason": "Missing fields from trust_remote_code AutoConfig return values is not the same as an unrecognized-model/config.json model_type error." + }, + { + "left": "issue:44964", + "right": "issue:45600", + "accept": false, + "reason": "Model loading failure for Phi-4 multimodal is unrelated to stale auto_mappings references to removed Sam3LiteText configs." + }, + { + "left": "issue:44112", + "right": "issue:44964", + "accept": false, + "reason": "A CI test regression on GraniteSpeech device override is unrelated to loading Phi-4 multimodal models." + }, + { + "left": "issue:42907", + "right": "issue:43441", + "accept": false, + "reason": "Saving dequantized Ministral/Devstral models and FlashAttention loading failures are separate model-specific bugs." + }, + { + "left": "issue:43854", + "right": "issue:44964", + "accept": false, + "reason": "Unit-test load failure for GLM-4.7-Flash and Phi-4 multimodal loading failure are different models and code paths." + }, + { + "left": "issue:44016", + "right": "issue:44062", + "accept": false, + "reason": "Notebook syntax error and AddedToken keyword collision are unrelated issues." + }, + { + "left": "issue:43976", + "right": "issue:44987", + "accept": false, + "reason": "Python version compatibility breakage is not the same as a specific model loading failure on 5.1.0+." + }, + { + "left": "issue:43295", + "right": "issue:45335", + "accept": false, + "reason": "Processor/tokenizer regression and t5gemma resize_token_embeddings bug affect different APIs and model components." + }, + { + "left": "issue:43673", + "right": "issue:44336", + "accept": false, + "reason": "Generation cache missing during chunked_prefill is unrelated to ANSI codes emitted by loading_report." + }, + { + "left": "issue:43531", + "right": "issue:45600", + "accept": false, + "reason": "Qwen3-MoE sliding_window behavior and stale auto_mappings references are different bugs." + }, + { + "left": "issue:43824", + "right": "issue:44933", + "accept": false, + "reason": "Qwen2.5-VL import/export failure is not the same as a nonexistent import from image_utils." + }, + { + "left": "issue:44496", + "right": "issue:45042", + "accept": false, + "reason": "Unrecognized model/config error is different from PIL image processors incorrectly requiring torchvision." + }, + { + "left": "issue:43526", + "right": "issue:43901", + "accept": false, + "reason": "BeitImageProcessorFast label reduction bug is unrelated to TextClassificationPipeline docs mismatch." + }, + { + "left": "issue:43299", + "right": "issue:45216", + "accept": false, + "reason": "Qwen3VL MoE loading breakage is a different model family and failure mode from Qwen3.5 save_pretrained corruption." + }, + { + "left": "issue:44075", + "right": "issue:44368", + "accept": false, + "reason": "SGD optimizer arguments not being used is unrelated to a tie_word_embeddings warning in Qwen3.5 fine-tuning." + }, + { + "left": "issue:45335", + "right": "issue:45584", + "accept": false, + "reason": "t5gemma embedding resize bug is unrelated to Whisper empty transcription generation after token alignment." + }, + { + "left": "issue:43976", + "right": "issue:44279", + "accept": false, + "reason": "Python compatibility and dependency issues are both environment-related, but they are not the same concrete failure." + }, + { + "left": "issue:43381", + "right": "issue:45593", + "accept": false, + "reason": "Gradient checkpointing in eval mode and D-FINE auxiliary losses with denoising off are unrelated training behaviors." + }, + { + "left": "issue:43299", + "right": "issue:45357", + "accept": false, + "reason": "Both are Qwen3.5-related regressions, but one is loading failure and the other is incorrect save_pretrained visual encoder keys." + }, + { + "left": "issue:43901", + "right": "issue:44496", + "accept": false, + "reason": "Docs mentioning return_all_scores and a model config loading error are not the same bug." + }, + { + "left": "issue:44496", + "right": "issue:45759", + "accept": false, + "reason": "Unrecognized model config loading failure is different from AutoModelForCausalLM not unwrapping text_config for composite multimodal configs." + }, + { + "left": "issue:44351", + "right": "issue:45278", + "accept": false, + "reason": "Missing HybridCache import is a specific symbol export issue, not the broad import-error cluster described by 45278." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 32, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:33290", + "issue:38175", + "issue:39692", + "issue:41950", + "issue:42994", + "issue:43065", + "issue:43295", + "issue:43299", + "issue:43317", + "issue:43441", + "issue:43450", + "issue:43479", + "issue:43525", + "issue:43577", + "issue:43650", + "issue:43854", + "issue:43901", + "issue:43976", + "issue:44016", + "issue:44242", + "issue:44261", + "issue:44263", + "issue:44279", + "issue:44291", + "issue:44297", + "issue:44462", + "issue:44466", + "issue:44521", + "issue:44617", + "issue:44683", + "issue:44779", + "issue:44898", + "issue:45020", + "issue:45081", + "issue:45200", + "issue:45216", + "issue:45276", + "issue:45290", + "issue:45325", + "issue:45405", + "issue:45542", + "issue:45600", + "issue:45656", + "issue:45712", + "issue:45812", + "issue:45823", + "issue:45834", + "issue:45907" + ], + "soft_pairs": [ + "issue:44521|issue:45325", + "issue:41950|issue:45020", + "issue:43976|issue:45405", + "issue:44242|issue:45200", + "issue:43479|issue:43525", + "issue:44779|issue:45812", + "issue:41950|issue:45823", + "issue:38175|issue:44617", + "issue:43577|issue:45216", + "issue:44683|issue:45542", + "issue:39692|issue:43901", + "issue:43299|issue:43854", + "issue:44279|issue:45405", + "issue:41950|issue:43650", + "issue:43295|issue:44779", + "issue:44462|issue:45081", + "issue:44297|issue:44462", + "issue:45081|issue:45907", + "issue:44016|issue:44291", + "issue:43441|issue:45834", + "issue:33290|issue:45656", + "issue:45600|issue:45712", + "issue:43450|issue:44279", + "issue:43065|issue:44898", + "issue:42994|issue:43317", + "issue:44466|issue:45276", + "issue:44261|issue:44263", + "issue:43650|issue:45290" + ], + "prepared_review_unit_hash": "b49b549da53e24330d425c68000b5931dc6369d387d8d5fa41c09812b82d64cf", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 28, + "serialized_chars": 36117, + "estimated_input_tokens": 9030, + "estimated_eval_tokens": 18316 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is largely a collection of unrelated Transformers issues with superficial similarity around regressions, tokenizers, multimodal models, and deepspeed. I found no strong duplicate pairs to merge; the soft matches are mostly same-subsystem, not same-bug matches.", + "confidence": 0.91, + "canonical_issue_reason": "issue:45020 is the broadest representative report here: it describes a recent-version regression affecting model loading/remote code and best matches the cluster\u2019s overall \u2018v5 breakage\u2019 theme.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45020 is the most suitable anchor because it is broad, open, and discussion-heavy, while the other issues are narrower, model-specific failures.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44521", + "right": "issue:45325", + "accept": false, + "reason": "Both are multimodal, but one is a chat-template masking bug and the other is a RoPE/position-id scaling issue for Qwen2.5-VL. Different code paths and symptoms." + }, + { + "left": "issue:41950", + "right": "issue:45020", + "accept": false, + "reason": "Video-classification processor lookup and remote-code model breakage are unrelated failures." + }, + { + "left": "issue:43976", + "right": "issue:45405", + "accept": false, + "reason": "Python version support failure and a PEFT minimum-version bump are not the same bug." + }, + { + "left": "issue:44242", + "right": "issue:45200", + "accept": false, + "reason": "Load-balancing loss gating and Gemma4 mm_token_type_ids defaults are different model behaviors." + }, + { + "left": "issue:43479", + "right": "issue:43525", + "accept": false, + "reason": "Different configs and different failures: default multimodal config initialization vs missing pad_token_id on Llama4Config." + }, + { + "left": "issue:44779", + "right": "issue:45812", + "accept": false, + "reason": "Both are tokenizer regressions, but they target different model families and likely different tokenizer code paths." + }, + { + "left": "issue:41950", + "right": "issue:45823", + "accept": false, + "reason": "Processor selection for video classification is unrelated to Gemma4 device placement during forward." + }, + { + "left": "issue:38175", + "right": "issue:44617", + "accept": false, + "reason": "Zero probabilities in SigLIP2 and CUDA OOM in Sam3Video are unrelated issues." + }, + { + "left": "issue:43577", + "right": "issue:45216", + "accept": false, + "reason": "BLIP2 dtype propagation and Qwen3.5 save_pretrained serialization are distinct bugs." + }, + { + "left": "issue:44683", + "right": "issue:45542", + "accept": false, + "reason": "Compiled flex_attention on torch 2.9 and tensorboard/TensorFlow backend installation are unrelated." + }, + { + "left": "issue:39692", + "right": "issue:43901", + "accept": false, + "reason": "Both involve docs, but one is a SigLIP2 example error and the other is a pipeline docs mismatch." + }, + { + "left": "issue:43299", + "right": "issue:43854", + "accept": false, + "reason": "Different model families and different loading failures; not the same underlying regression." + }, + { + "left": "issue:44279", + "right": "issue:45405", + "accept": false, + "reason": "A generic dependency issue is not the same as a specific unreleased PEFT version bump." + }, + { + "left": "issue:41950", + "right": "issue:43650", + "accept": false, + "reason": "Unrelated: video pipeline processor lookup vs an unrelated placeholder issue." + }, + { + "left": "issue:43295", + "right": "issue:44779", + "accept": false, + "reason": "Processor/tokenizer regression in v4.57.5 is separate from Deepseek tokenizer misbehavior in v5." + }, + { + "left": "issue:44462", + "right": "issue:45081", + "accept": false, + "reason": "AutoTokenizer ignoring tokenizer.json and Mistral regex patch crashing are different tokenizer-loading bugs." + }, + { + "left": "issue:44297", + "right": "issue:44462", + "accept": false, + "reason": "Tokenizer class serialization mismatch is not the same as repository tokenizer.json being ignored." + }, + { + "left": "issue:45081", + "right": "issue:45907", + "accept": false, + "reason": "Mistral tokenizer loading and list_repo_files revision forwarding are unrelated." + }, + { + "left": "issue:44016", + "right": "issue:44291", + "accept": false, + "reason": "Notebook syntax error and init_empty_weights argument mismatch are unrelated." + }, + { + "left": "issue:43441", + "right": "issue:45834", + "accept": false, + "reason": "FlashAttention failure on Ministral-3 and OCR index error in Kosmos2.5 are different model-specific bugs." + }, + { + "left": "issue:33290", + "right": "issue:45656", + "accept": false, + "reason": "Both mention deepspeed/optimizers, but OOM with Adafactor and double optimizer.step calls are different issues." + }, + { + "left": "issue:45600", + "right": "issue:45712", + "accept": false, + "reason": "Both are repo hygiene/CI problems, but they concern different files and failure modes." + }, + { + "left": "issue:43450", + "right": "issue:44279", + "accept": false, + "reason": "Video processor batching shape bug and a dependency issue are unrelated." + }, + { + "left": "issue:43065", + "right": "issue:44898", + "accept": false, + "reason": "Different vision-model bugs: dummy Conv2d in Sam3PixelDecoder vs Perceiver non-default resolution failure." + }, + { + "left": "issue:42994", + "right": "issue:43317", + "accept": false, + "reason": "Quantized model saving failure and dequantized load/offload failure are distinct stages and bugs." + }, + { + "left": "issue:44466", + "right": "issue:45276", + "accept": false, + "reason": "Serialization of lm_head.weight and resize_token_embeddings not updating Gemma4 outputs are separate problems." + }, + { + "left": "issue:44261", + "right": "issue:44263", + "accept": false, + "reason": "MLA layernorm epsilon mismatch and GlmMoeDsaIndexer split-return handling are unrelated." + }, + { + "left": "issue:43650", + "right": "issue:45290", + "accept": false, + "reason": "Placeholder issue and apply_chat_template crash with tool-call assistant messages are unrelated." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 33, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:41628", + "issue:42175", + "issue:42491", + "issue:43122", + "issue:43262", + "issue:43299", + "issue:43441", + "issue:43450", + "issue:43525", + "issue:43643", + "issue:43650", + "issue:43653", + "issue:43673", + "issue:43723", + "issue:43761", + "issue:43784", + "issue:43824", + "issue:43881", + "issue:43901", + "issue:43976", + "issue:43994", + "issue:44016", + "issue:44162", + "issue:44188", + "issue:44242", + "issue:44246", + "issue:44295", + "issue:44297", + "issue:44336", + "issue:44361", + "issue:44373", + "issue:44521", + "issue:44589", + "issue:44617", + "issue:44623", + "issue:44671", + "issue:44683", + "issue:44749", + "issue:44779", + "issue:44987", + "issue:45362", + "issue:45397", + "issue:45542", + "issue:45561", + "issue:45563", + "issue:45593", + "issue:45823" + ], + "soft_pairs": [ + "issue:44623|issue:45397", + "issue:43441|issue:44683", + "issue:44623|issue:44749", + "issue:43653|issue:45593", + "issue:43262|issue:44521", + "issue:43723|issue:44162", + "issue:44016|issue:45823", + "issue:43673|issue:43761", + "issue:44361|issue:45593", + "issue:42175|issue:43976", + "issue:44671|issue:44779", + "issue:43784|issue:45542", + "issue:43901|issue:43994", + "issue:43122|issue:43650", + "issue:44987|issue:45542", + "issue:43824|issue:43976", + "issue:44297|issue:45397", + "issue:44188|issue:44336", + "issue:43525|issue:43881", + "issue:43525|issue:43643", + "issue:43450|issue:44617", + "issue:44589|issue:45362", + "issue:41628|issue:43976", + "issue:45561|issue:45563", + "issue:43901|issue:44246", + "issue:42491|issue:43299", + "issue:44336|issue:44373", + "issue:44242|issue:45593", + "issue:44295|issue:45397" + ], + "prepared_review_unit_hash": "89f96f0701692593d902868c841374adc4472e841bb338ecc00da64aff6446f1", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 29, + "serialized_chars": 35460, + "estimated_input_tokens": 8865, + "estimated_eval_tokens": 17986 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is mostly a loose embedding match around Transformers v5-era regressions, but the paired items are on different code paths (tokenizers, model loading, chat templates, attention kernels, backend packaging, docs, etc.). I do not see any soft pair that is close enough to be treated as the same underlying bug/change.", + "confidence": 0.91, + "canonical_issue_reason": "issue:44521 is the most central user-facing bug in this set: it is broad, clearly described, open, and has the most obvious potential to attract/collect related multimodal chat-template regressions.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:44521 is the best representative issue here because it is concrete, high-signal, and more likely than the others to serve as a meaningful anchor for nearby regressions.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44623", + "right": "issue:45397", + "accept": false, + "reason": "Different failures: saving processor artifacts vs loading a model under zero3; no shared concrete code-path." + }, + { + "left": "issue:43441", + "right": "issue:44683", + "accept": false, + "reason": "Both mention attention backends, but one is FlashAttention in Ministral-3 and the other is compiled flex_attention on torch>=2.9; distinct bugs." + }, + { + "left": "issue:44623", + "right": "issue:44749", + "accept": false, + "reason": "Processor save_pretrained missing files and tokenizer throughput regression are unrelated." + }, + { + "left": "issue:43653", + "right": "issue:45593", + "accept": false, + "reason": "BigBirdTokenizer special-token registration is unrelated to D-FINE auxiliary loss logic." + }, + { + "left": "issue:43262", + "right": "issue:44521", + "accept": false, + "reason": "Same API family, but different bug: audio sampling-rate default vs multimodal assistant mask generation." + }, + { + "left": "issue:43723", + "right": "issue:44162", + "accept": false, + "reason": "Tokenizer loading regression and ESM2 model breakage are different issues." + }, + { + "left": "issue:44016", + "right": "issue:45823", + "accept": false, + "reason": "Notebook syntax error is not the same as Gemma4 device-map forward mismatch." + }, + { + "left": "issue:43673", + "right": "issue:43761", + "accept": false, + "reason": "Generation cache/chunked_prefill failure and CLIPVision hidden_states regression affect different modules and behaviors." + }, + { + "left": "issue:44361", + "right": "issue:45593", + "accept": false, + "reason": "Tokenizer task AttributeError and D-FINE auxiliary loss omission are unrelated." + }, + { + "left": "issue:42175", + "right": "issue:43976", + "accept": false, + "reason": "Backend packaging/install issue and Python version compatibility issue are different root causes." + }, + { + "left": "issue:44671", + "right": "issue:44779", + "accept": false, + "reason": "Different model families and different symptoms: CamemBERT MLM predictions vs Deepseek tokenization." + }, + { + "left": "issue:43784", + "right": "issue:45542", + "accept": false, + "reason": "Both are import/backend-related, but one is a missing nn import in sentence-transformers integration and the other is a missing TensorFlow backend when only tensorboard is installed." + }, + { + "left": "issue:43901", + "right": "issue:43994", + "accept": false, + "reason": "Docs mismatch about return_all_scores is not the same as incorrect SigLIP2 inference outputs." + }, + { + "left": "issue:43122", + "right": "issue:43650", + "accept": false, + "reason": "Tokenizer output drift and a nonspecific 'ADD THE DATA' issue do not identify the same bug." + }, + { + "left": "issue:44987", + "right": "issue:45542", + "accept": false, + "reason": "Model loading failure for a specific repo and TensorFlow backend packaging issue are unrelated." + }, + { + "left": "issue:43824", + "right": "issue:43976", + "accept": false, + "reason": "Qwen2_5_VL import failure and Python 3.9/3.10 compatibility are different failure modes." + }, + { + "left": "issue:44297", + "right": "issue:45397", + "accept": false, + "reason": "Tokenizer config serialization mismatch and Gemma4 loading under zero3 are different code paths." + }, + { + "left": "issue:44188", + "right": "issue:44336", + "accept": false, + "reason": "Attention-kernel divergence under torch.compile is unrelated to ANSI codes in loading_report." + }, + { + "left": "issue:43525", + "right": "issue:43881", + "accept": false, + "reason": "Llama4Config missing pad_token_id and glm-4v-9b loading failure are model-specific and not the same bug." + }, + { + "left": "issue:43525", + "right": "issue:43643", + "accept": false, + "reason": "One is a specific config attribute absence; the other is a generic trust_remote_code field-loss issue. Similar theme, different defect." + }, + { + "left": "issue:43450", + "right": "issue:44617", + "accept": false, + "reason": "Incorrect batched video processor shapes and Sam3Video CUDA OOM are not the same underlying problem." + }, + { + "left": "issue:44589", + "right": "issue:45362", + "accept": false, + "reason": "Float8 storage lookup failure and Qwen3.5 chat crash are different runtime errors." + }, + { + "left": "issue:41628", + "right": "issue:43976", + "accept": false, + "reason": "Importing AutoImageProcessor and Python version support are unrelated." + }, + { + "left": "issue:45561", + "right": "issue:45563", + "accept": false, + "reason": "pytest-xdist file race and stale generate() warning are unrelated test/runtime issues." + }, + { + "left": "issue:43901", + "right": "issue:44246", + "accept": false, + "reason": "Documentation note about return_all_scores and slow import behavior are different concerns." + }, + { + "left": "issue:42491", + "right": "issue:43299", + "accept": false, + "reason": "Both involve Qwen3 MoE, but one is LoRA compatibility across versions and the other is base-model loading regression; too different to merge as one fix." + }, + { + "left": "issue:44336", + "right": "issue:44373", + "accept": false, + "reason": "ANSI output formatting and a wrong docstring are not the same bug." + }, + { + "left": "issue:44242", + "right": "issue:45593", + "accept": false, + "reason": "Load-balancing loss routing logic and D-FINE auxiliary-loss behavior are unrelated." + }, + { + "left": "issue:44295", + "right": "issue:45397", + "accept": false, + "reason": "Buffer registration/readback error and Gemma4 zero3 loading failure are different defects." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 34, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:36246", + "issue:41950", + "issue:42757", + "issue:42886", + "issue:43065", + "issue:43232", + "issue:43257", + "issue:43262", + "issue:43352", + "issue:43408", + "issue:43526", + "issue:43531", + "issue:43596", + "issue:43673", + "issue:43901", + "issue:44060", + "issue:44077", + "issue:44162", + "issue:44220", + "issue:44242", + "issue:44246", + "issue:44295", + "issue:44336", + "issue:44351", + "issue:44485", + "issue:44496", + "issue:44561", + "issue:44749", + "issue:44779", + "issue:44857", + "issue:44871", + "issue:44877", + "issue:44987", + "issue:44995", + "issue:45003", + "issue:45137", + "issue:45200", + "issue:45229", + "issue:45250", + "issue:45292", + "issue:45362", + "issue:45375", + "issue:45459", + "issue:45561", + "issue:45584", + "issue:45701", + "issue:45759" + ], + "soft_pairs": [ + "issue:41950|issue:44220", + "issue:44246|issue:44496", + "issue:43531|issue:45362", + "issue:36246|issue:42757", + "issue:43352|issue:45250", + "issue:36246|issue:44351", + "issue:44336|issue:44485", + "issue:43526|issue:45003", + "issue:45003|issue:45561", + "issue:43673|issue:44561", + "issue:44246|issue:44871", + "issue:44242|issue:44857", + "issue:44162|issue:45200", + "issue:44295|issue:44749", + "issue:36246|issue:43596", + "issue:44877|issue:45375", + "issue:44779|issue:45292", + "issue:45292|issue:45584", + "issue:43901|issue:45003", + "issue:45292|issue:45701", + "issue:43232|issue:44336", + "issue:42886|issue:45229", + "issue:43065|issue:43408", + "issue:44987|issue:45459", + "issue:44496|issue:45003", + "issue:45375|issue:45759", + "issue:43257|issue:45137", + "issue:43262|issue:44246", + "issue:44995|issue:45200", + "issue:44060|issue:44077" + ], + "prepared_review_unit_hash": "70abf0262e597b254f4f0fd35406be8df6929df57bd75646db8323e6ed6ffa78", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 30, + "serialized_chars": 35958, + "estimated_input_tokens": 8990, + "estimated_eval_tokens": 18236 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "No soft pair looks like the same underlying bug or change. The set is heterogeneous across unrelated areas (pipelines, tokenizer/config loading, generation, DeepSpeed, and model-specific issues), so none should be deduplicated.", + "confidence": 0.94, + "canonical_issue_reason": "No single canonical issue fits this cluster: the candidates cover unrelated defects and no pair looks mergeable as the same bug.", + "canonical_pr_reason": null, + "best_issue_reason": "No global best issue for deduplication; the list is too mixed to anchor a canonical representative.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:41950", + "right": "issue:44220", + "accept": false, + "reason": "Different subsystems and symptoms: video-classification pipeline processor lookup vs audio fbank feature extraction." + }, + { + "left": "issue:44246", + "right": "issue:44496", + "accept": false, + "reason": "Import performance issue is unrelated to model config parsing failure for an unrecognized model." + }, + { + "left": "issue:43531", + "right": "issue:45362", + "accept": false, + "reason": "Both involve Qwen, but the bugs differ: sliding-window handling vs chat crash on Qwen3.5-35B." + }, + { + "left": "issue:36246", + "right": "issue:42757", + "accept": false, + "reason": "Distinct import errors in different packages: Qwen2.5-VL image processor vs huggingface_hub offline helper." + }, + { + "left": "issue:43352", + "right": "issue:45250", + "accept": false, + "reason": "One is a model-specific Flash Attention support error; the other is only a broad feature title with no matching code path shown." + }, + { + "left": "issue:36246", + "right": "issue:44351", + "accept": false, + "reason": "Different missing imports from different modules; no shared code-path or fix." + }, + { + "left": "issue:44336", + "right": "issue:44485", + "accept": false, + "reason": "ANSI color output in loading reports is unrelated to GLM-5 RoPE implementation." + }, + { + "left": "issue:43526", + "right": "issue:45003", + "accept": false, + "reason": "BEiT image processor label reduction bug is unrelated to unsafe sys.modules access in modeling_utils." + }, + { + "left": "issue:45003", + "right": "issue:45561", + "accept": false, + "reason": "Unsafe module lookup and xdist file race are separate testing/runtime problems." + }, + { + "left": "issue:43673", + "right": "issue:44561", + "accept": false, + "reason": "Chunked-prefill cache bug is unrelated to removal of is_torch_fx_available breaking trust_remote_code models." + }, + { + "left": "issue:44246", + "right": "issue:44871", + "accept": false, + "reason": "Import slowdown has no concrete relation to Gemma eos_token_id configuration mismatch." + }, + { + "left": "issue:44242", + "right": "issue:44857", + "accept": false, + "reason": "Load-balancing loss omission and LWDETR AMP crash are different MoE vs vision-loss bugs." + }, + { + "left": "issue:44162", + "right": "issue:45200", + "accept": false, + "reason": "ESM2 breakage and Gemma 4 mm_token_type_ids defaulting are unrelated model-specific issues." + }, + { + "left": "issue:44295", + "right": "issue:44749", + "accept": false, + "reason": "Position_ids buffer access error is unrelated to the reported tokenization slowdown." + }, + { + "left": "issue:36246", + "right": "issue:43596", + "accept": false, + "reason": "Qwen2.5-VL import failure and DeepSpeed ZeRO3 deque IndexError are unrelated." + }, + { + "left": "issue:44877", + "right": "issue:45375", + "accept": false, + "reason": "Strict config loading for granite_speech is not the same as missing Qwen3_5MoeVisionConfig field handling." + }, + { + "left": "issue:44779", + "right": "issue:45292", + "accept": false, + "reason": "Tokenizer regression and resize_token_embeddings output embedding mismatch are different bugs." + }, + { + "left": "issue:45292", + "right": "issue:45584", + "accept": false, + "reason": "Embedding resize behavior is unrelated to Whisper empty-transcription generation failure." + }, + { + "left": "issue:43901", + "right": "issue:45003", + "accept": false, + "reason": "Docs mismatch for return_all_scores is unrelated to sys.modules access." + }, + { + "left": "issue:45292", + "right": "issue:45701", + "accept": false, + "reason": "resize_token_embeddings/output embeddings bug is not the same as version-dependent tokenization changes." + }, + { + "left": "issue:43232", + "right": "issue:44336", + "accept": false, + "reason": "Generation kwargs update after sync_gpus is unrelated to ANSI codes in loading_report." + }, + { + "left": "issue:42886", + "right": "issue:45229", + "accept": false, + "reason": "Offline tokenizer cache loading bug is unrelated to Gemma4 multi-GPU OOM." + }, + { + "left": "issue:43065", + "right": "issue:43408", + "accept": false, + "reason": "Dummy Conv2d in Sam3PixelDecoder and sam3_video/sam3_tracker config mismatch are different SAM3 issues." + }, + { + "left": "issue:44987", + "right": "issue:45459", + "accept": false, + "reason": "Model loading failure for physical-intelligence/fast is unrelated to tokenizer exception masking when protobuf is absent." + }, + { + "left": "issue:44496", + "right": "issue:45003", + "accept": false, + "reason": "Unrecognized model/config error is unrelated to unsafe sys.modules access." + }, + { + "left": "issue:45375", + "right": "issue:45759", + "accept": false, + "reason": "Missing config field in Qwen3_5MoeVisionConfig is not the same as AutoModelForCausalLM unwrapping text_config." + }, + { + "left": "issue:43257", + "right": "issue:45137", + "accept": false, + "reason": "Qwen3 MOE weight conversion under accelerate+deepspeed and ZeRO3 deque popping are different failure modes." + }, + { + "left": "issue:43262", + "right": "issue:44246", + "accept": false, + "reason": "Audio chat-template sample-rate default bug is unrelated to import performance." + }, + { + "left": "issue:44995", + "right": "issue:45200", + "accept": false, + "reason": "GlmMoeDsa stale indexer cache crash does not match Gemma4 token-type-id defaults." + }, + { + "left": "issue:44060", + "right": "issue:44077", + "accept": false, + "reason": "Qwen3.5 tied-weights warning and patchtsmixer optional post_init are unrelated model/config issues." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 35, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:28282", + "issue:30990", + "issue:41628", + "issue:42371", + "issue:42994", + "issue:43232", + "issue:43475", + "issue:43482", + "issue:43525", + "issue:43540", + "issue:43643", + "issue:43646", + "issue:43653", + "issue:43673", + "issue:43931", + "issue:44038", + "issue:44162", + "issue:44188", + "issue:44297", + "issue:44361", + "issue:44368", + "issue:44466", + "issue:44623", + "issue:44661", + "issue:44734", + "issue:44749", + "issue:44869", + "issue:44933", + "issue:44977", + "issue:44995", + "issue:45081", + "issue:45092", + "issue:45103", + "issue:45137", + "issue:45230", + "issue:45278", + "issue:45310", + "issue:45335", + "issue:45356", + "issue:45375", + "issue:45381", + "issue:45397", + "issue:45459", + "issue:45478", + "issue:45563", + "issue:45588", + "issue:45600", + "issue:45656" + ], + "soft_pairs": [ + "issue:43475|issue:43540", + "issue:43653|issue:44162", + "issue:44466|issue:45335", + "issue:45397|issue:45656", + "issue:44995|issue:45588", + "issue:44368|issue:45563", + "issue:44368|issue:44977", + "issue:43673|issue:44734", + "issue:45137|issue:45600", + "issue:43931|issue:44368", + "issue:44933|issue:45459", + "issue:44869|issue:45356", + "issue:43232|issue:43673", + "issue:44297|issue:44623", + "issue:43673|issue:45278", + "issue:43646|issue:44162", + "issue:44361|issue:44749", + "issue:44623|issue:45656", + "issue:44749|issue:45656", + "issue:44749|issue:45310", + "issue:28282|issue:30990", + "issue:42994|issue:43482", + "issue:45230|issue:45397", + "issue:42371|issue:45092", + "issue:43475|issue:43643", + "issue:43475|issue:45381", + "issue:44038|issue:44623", + "issue:41628|issue:44933", + "issue:45081|issue:45103", + "issue:43475|issue:44661", + "issue:42994|issue:43525", + "issue:45375|issue:45478", + "issue:43232|issue:44188" + ], + "prepared_review_unit_hash": "589c4b661defa5488cb3fea9cb22746fb1831c6b1d232049676fe27c1890884f", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 33, + "serialized_chars": 37588, + "estimated_input_tokens": 9397, + "estimated_eval_tokens": 19050 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The set is mostly a grab bag of unrelated regression reports across tokenizers, generation, model loading, deepspeed, vision, and import/export paths. I did not find any soft pair that looks like the same concrete bug fix target, so no duplicate merges are warranted.", + "confidence": 0.79, + "canonical_issue_reason": "issue:45278 is the broadest current umbrella report in the set: it targets upgrade-related import regressions across many cases and is open, making it the best central triage anchor.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45278 is the best single issue to keep as the representative hub because it is broad, current, and likely to accumulate related import-regression reports.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43475", + "right": "issue:43540", + "accept": false, + "reason": "Different models and failure modes: SAM 3 video missing an encoder output attribute vs Qwen3OmniMoe video-input processing ValueError." + }, + { + "left": "issue:43653", + "right": "issue:44162", + "accept": false, + "reason": "One is a BigBirdTokenizer special-token registration bug; the other is an ESM2 breakage report affecting workflows, not the same code path." + }, + { + "left": "issue:44466", + "right": "issue:45335", + "accept": false, + "reason": "Both involve weight/embedding serialization, but one is tied lm_head serialization by device and the other is resize_token_embeddings not updating decoder.embed_tokens; different bugs." + }, + { + "left": "issue:45397", + "right": "issue:45656", + "accept": false, + "reason": "Both mention DeepSpeed, but one is a gemma-4 from_pretrained/ZeRO3 issue and the other is double optimizer stepping; unrelated code paths." + }, + { + "left": "issue:44995", + "right": "issue:45588", + "accept": false, + "reason": "A second-forward stale indexer cache crash is not the same as a flash-attention crash on s_aux=None." + }, + { + "left": "issue:44368", + "right": "issue:45563", + "accept": false, + "reason": "One is a tie_word_embeddings warning during fine-tuning; the other is a stale num_return_sequences warning in paged generate()." + }, + { + "left": "issue:44368", + "right": "issue:44977", + "accept": false, + "reason": "A warning about tie_word_embeddings is unrelated to Qwen3.5 generation failures under flash-attention." + }, + { + "left": "issue:43673", + "right": "issue:44734", + "accept": false, + "reason": "Both touch KV/cache behavior, but one is missing GenerationMixin cache in chunked_prefill and the other is wrong tensor indexing during /v1/responses continuation." + }, + { + "left": "issue:45137", + "right": "issue:45600", + "accept": false, + "reason": "DeepSpeed ZeRO3 deque underflow is unrelated to stale auto_mappings references to removed Sam3LiteText configs." + }, + { + "left": "issue:43931", + "right": "issue:44368", + "accept": false, + "reason": "Qwen3-VL weight-shape mismatch is unrelated to a config warning about tie_word_embeddings." + }, + { + "left": "issue:44933", + "right": "issue:45459", + "accept": false, + "reason": "A missing import from image_utils is not the same as protobuf decode errors being masked by except import_protobuf_decode_error()." + }, + { + "left": "issue:44869", + "right": "issue:45356", + "accept": false, + "reason": "Whisper trailing-replacement-character timestamp decoding and Kimi-K2.5 codec/fix_mistral_regex regression are different tokenizer issues." + }, + { + "left": "issue:43232", + "right": "issue:43673", + "accept": false, + "reason": "Both are generation-related, but one is sync_gpus/update_model_kwargs logic and the other is v5 chunked_prefill cache availability." + }, + { + "left": "issue:44297", + "right": "issue:44623", + "accept": false, + "reason": "tokenizer.save_pretrained writing the wrong tokenizer_class is a different save-path bug from processor.save_pretrained missing files." + }, + { + "left": "issue:43673", + "right": "issue:45278", + "accept": false, + "reason": "Chunked_prefill cache handling is not the same as broad import errors after upgrading from 4.57.0 to 5.5.0." + }, + { + "left": "issue:43646", + "right": "issue:44162", + "accept": false, + "reason": "Custom model initialization breakage in v5 is too broad to equate with an ESM2-specific workflow regression." + }, + { + "left": "issue:44361", + "right": "issue:44749", + "accept": false, + "reason": "MLukeTokenizer AttributeError on tasks is unrelated to performance slowdown in filtering after upgrading Transformers." + }, + { + "left": "issue:44623", + "right": "issue:45656", + "accept": false, + "reason": "Missing files from processor.save_pretrained is unrelated to DeepSpeed causing the optimizer step to run twice." + }, + { + "left": "issue:44749", + "right": "issue:45656", + "accept": false, + "reason": "A performance regression in data filtering and a DeepSpeed double-step bug are not the same underlying issue." + }, + { + "left": "issue:44749", + "right": "issue:45310", + "accept": false, + "reason": "Slow filtering after upgrade is unrelated to Qwen3.5 Moe from_pretrained errors." + }, + { + "left": "issue:28282", + "right": "issue:30990", + "accept": false, + "reason": "PyTorch-missing ImportError and Sentence Transformers hanging on load are different environment/loading problems." + }, + { + "left": "issue:42994", + "right": "issue:43482", + "accept": false, + "reason": "Quantized model saving failure is not the same as Qwen2.5-GGUF loading failure in v5." + }, + { + "left": "issue:45230", + "right": "issue:45397", + "accept": false, + "reason": "A placeholder bug report does not identify the same from_pretrained/ZeRO3 issue as gemma-4." + }, + { + "left": "issue:42371", + "right": "issue:45092", + "accept": false, + "reason": "TF32 API-setting guidance is unrelated to old InternVL2 checkpoints breaking under meta initialization." + }, + { + "left": "issue:43475", + "right": "issue:43643", + "accept": false, + "reason": "SAM 3 video encoder output missing an attribute is unrelated to AutoConfig returning missing fields under trust_remote_code=True." + }, + { + "left": "issue:43475", + "right": "issue:45381", + "accept": false, + "reason": "Different vision/video regressions in different models: SAM 3 output attribute error vs Qwen2.5-VL video position IDs." + }, + { + "left": "issue:44038", + "right": "issue:44623", + "accept": false, + "reason": "Qwen3-VL-Moe from_pretrained breakage and processor.save_pretrained missing files are separate bugs." + }, + { + "left": "issue:41628", + "right": "issue:44933", + "accept": false, + "reason": "An AutoImageProcessor import error and a missing image_utils import are related only at a broad import level, not the same concrete bug." + }, + { + "left": "issue:45081", + "right": "issue:45103", + "accept": false, + "reason": "Mistral regex patch crashing on tokenizer backend access is unrelated to auto_docstring crashing on future annotations." + }, + { + "left": "issue:43475", + "right": "issue:44661", + "accept": false, + "reason": "SAM 3 vision encoder output and add-new-model-like failing inside TOKENIZER_MAPPING_NAMES are different subsystems and failures." + }, + { + "left": "issue:42994", + "right": "issue:43525", + "accept": false, + "reason": "Quantized model saving and Llama4Config missing pad_token_id are unrelated." + }, + { + "left": "issue:45375", + "right": "issue:45478", + "accept": false, + "reason": "Missing deepstack_visual_indexes in a vision config is not the same as Qwen3.5 Moe from_pretrained errors." + }, + { + "left": "issue:43232", + "right": "issue:44188", + "accept": false, + "reason": "GenerationMixin cache handling after sync_gpus is unrelated to diverging attention kernels under torch.compile." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 36, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:30990", + "issue:35141", + "issue:36010", + "issue:41628", + "issue:42175", + "issue:42222", + "issue:42673", + "issue:42831", + "issue:42898", + "issue:42994", + "issue:43232", + "issue:43278", + "issue:43377", + "issue:43479", + "issue:43504", + "issue:43526", + "issue:43673", + "issue:43704", + "issue:43784", + "issue:43824", + "issue:43873", + "issue:43976", + "issue:44038", + "issue:44079", + "issue:44164", + "issue:44188", + "issue:44222", + "issue:44242", + "issue:44351", + "issue:44373", + "issue:44496", + "issue:44568", + "issue:44623", + "issue:44734", + "issue:44792", + "issue:44869", + "issue:44933", + "issue:44945", + "issue:44964", + "issue:45103", + "issue:45137", + "issue:45200", + "issue:45310", + "issue:45375", + "issue:45561", + "issue:45712", + "issue:45753", + "issue:45812" + ], + "soft_pairs": [ + "issue:42673|issue:43704", + "issue:45310|issue:45375", + "issue:43873|issue:44945", + "issue:41628|issue:44351", + "issue:42831|issue:44222", + "issue:44945|issue:45753", + "issue:44222|issue:45310", + "issue:43673|issue:44188", + "issue:36010|issue:41628", + "issue:42994|issue:44164", + "issue:42175|issue:45712", + "issue:42222|issue:45200", + "issue:43232|issue:44792", + "issue:44038|issue:45310", + "issue:43377|issue:44242", + "issue:43526|issue:45561", + "issue:43504|issue:45561", + "issue:44496|issue:45561", + "issue:44964|issue:45137", + "issue:42175|issue:44933", + "issue:42898|issue:44568", + "issue:30990|issue:35141", + "issue:43479|issue:44869", + "issue:43976|issue:44933", + "issue:43479|issue:44079", + "issue:44188|issue:44373", + "issue:43784|issue:43824", + "issue:44623|issue:45310", + "issue:42898|issue:45812", + "issue:43278|issue:45137", + "issue:44734|issue:45103" + ], + "prepared_review_unit_hash": "cac00b0d0602e6f1ddcd2530eeddf288650349207212cbda1a0e0d5a7e350041", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 31, + "serialized_chars": 36787, + "estimated_input_tokens": 9197, + "estimated_eval_tokens": 18650 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Mostly unrelated issue reports; the only clear duplicate pair is the identical Qwen3ForCausalLM VRAM leak report. No pull requests are present.", + "confidence": 0.94, + "canonical_issue_reason": "issue:42673 is the best canonical target because issue:43704 has the exact same title and describes the same Qwen3ForCausalLM VRAM leak in multiple dataloader threads.", + "canonical_pr_reason": null, + "best_issue_reason": "Same concrete bug as its duplicate: identical symptom, model, and multithreaded dataloader context, so 42673 is the strongest representative issue.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:42673", + "right": "issue:43704", + "accept": true, + "reason": "Exact same bug report and title; clearly the same Qwen3ForCausalLM VRAM leak issue." + }, + { + "left": "issue:45310", + "right": "issue:45375", + "accept": false, + "reason": "Both touch Qwen3.5/Moe config loading, but one is a generic from_pretrained error and the other is a specific missing-field/config-schema problem; not enough to call the same bug." + }, + { + "left": "issue:43873", + "right": "issue:44945", + "accept": false, + "reason": "Different failures: quantization/offloading vs incorrect output under pipeline parallelism." + }, + { + "left": "issue:41628", + "right": "issue:44351", + "accept": false, + "reason": "Both are import errors, but for different symbols and different root causes." + }, + { + "left": "issue:42831", + "right": "issue:44222", + "accept": false, + "reason": "Both mention FP8, but one is an accuracy issue and the other is a save_pretrained/MoE persistence bug." + }, + { + "left": "issue:44945", + "right": "issue:45753", + "accept": false, + "reason": "Pipeline-parallel incorrect output and a Qwen3_5 image-specific infinite loop are different concrete problems." + }, + { + "left": "issue:44222", + "right": "issue:45310", + "accept": false, + "reason": "FP8 save_pretrained and Qwen3.5 from_pretrained loading are unrelated bugs." + }, + { + "left": "issue:43673", + "right": "issue:44188", + "accept": false, + "reason": "Generation cache missing in chunked prefill and attention-kernel divergence under torch.compile are distinct code paths." + }, + { + "left": "issue:36010", + "right": "issue:41628", + "accept": false, + "reason": "Both are import failures, but for different APIs and unrelated breakages." + }, + { + "left": "issue:42994", + "right": "issue:44164", + "accept": false, + "reason": "Saving quantized models and handling extra_state in save/from_pretrained are separate issues." + }, + { + "left": "issue:42175", + "right": "issue:45712", + "accept": false, + "reason": "Packaging backend dependency issue vs leftover dummy-class hygiene issue; not the same bug." + }, + { + "left": "issue:42222", + "right": "issue:45200", + "accept": false, + "reason": "VitPose model breakage and Gemma 4 token-type defaults are unrelated." + }, + { + "left": "issue:43232", + "right": "issue:44792", + "accept": false, + "reason": "Generation kwargs update bug and a Janus image-generation test failure do not match." + }, + { + "left": "issue:44038", + "right": "issue:45310", + "accept": false, + "reason": "Same broad Qwen/Moe area, but different model variants and different failure modes/version regressions." + }, + { + "left": "issue:43377", + "right": "issue:44242", + "accept": false, + "reason": "MIMI padding-mask batching bug vs MoE load-balancing loss omission are different problems." + }, + { + "left": "issue:43526", + "right": "issue:45561", + "accept": false, + "reason": "BeitImageProcessorFast label reduction bug vs xdist race on captured_info.txt are unrelated." + }, + { + "left": "issue:43504", + "right": "issue:45561", + "accept": false, + "reason": "Legacy-field model loading issue vs pytest-xdist race are different bugs." + }, + { + "left": "issue:44496", + "right": "issue:45561", + "accept": false, + "reason": "Unrecognized model/config issue vs testing-utils race condition are unrelated." + }, + { + "left": "issue:44964", + "right": "issue:45137", + "accept": false, + "reason": "Phi-4 multimodal load failure and DeepSpeed ZeRO3 deque pop error are different code paths." + }, + { + "left": "issue:42175", + "right": "issue:44933", + "accept": false, + "reason": "Backend dependency install issue vs missing image_utils import are unrelated." + }, + { + "left": "issue:42898", + "right": "issue:44568", + "accept": false, + "reason": "Tokenizer cleanup behavior change and mDeBERTa special-token addition bug are different regressions." + }, + { + "left": "issue:30990", + "right": "issue:35141", + "accept": false, + "reason": "Sentence Transformers loading hang and embedding reinit after resize are unrelated." + }, + { + "left": "issue:43479", + "right": "issue:44869", + "accept": false, + "reason": "Multimodal config defaults vs Whisper timestamp decode crash are different issues." + }, + { + "left": "issue:43976", + "right": "issue:44933", + "accept": false, + "reason": "Python-version compatibility issue vs missing image_utils import are unrelated." + }, + { + "left": "issue:43479", + "right": "issue:44079", + "accept": false, + "reason": "Config defaulting bug and ModelOutput key assignment bug do not share the same underlying defect." + }, + { + "left": "issue:44188", + "right": "issue:44373", + "accept": false, + "reason": "Attention-kernel divergence and a wrong docstring are not the same bug." + }, + { + "left": "issue:43784", + "right": "issue:43824", + "accept": false, + "reason": "sentence-transformers import NameError and Qwen2_5_VL import error are unrelated." + }, + { + "left": "issue:44623", + "right": "issue:45310", + "accept": false, + "reason": "Missing processor.save_pretrained files and Qwen3.5 from_pretrained failure are different save/load problems." + }, + { + "left": "issue:42898", + "right": "issue:45812", + "accept": false, + "reason": "Tokenizer cleanup behavior change and Granite token-id regression are distinct tokenizer bugs." + }, + { + "left": "issue:43278", + "right": "issue:45137", + "accept": false, + "reason": "BF16-to-FP32 evaluation dtype change and ZeRO3 deque underflow are unrelated." + }, + { + "left": "issue:44734", + "right": "issue:45103", + "accept": false, + "reason": "KV-cache continuation crash in transformers serve and an auto_docstring annotations crash are unrelated." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "Grounded in the packet: the only clearly matching pair is the identical Qwen3ForCausalLM VRAM-leak issue titles, and the remaining soft-edge verdicts are conservative. No pull requests are present." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 37, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:22355", + "issue:28282", + "issue:30333", + "issue:33290", + "issue:34689", + "issue:35141", + "issue:36331", + "issue:41628", + "issue:42757", + "issue:42831", + "issue:42898", + "issue:43116", + "issue:43232", + "issue:43479", + "issue:43644", + "issue:43723", + "issue:43874", + "issue:44060", + "issue:44190", + "issue:44206", + "issue:44222", + "issue:44261", + "issue:44263", + "issue:44295", + "issue:44297", + "issue:44360", + "issue:44373", + "issue:44448", + "issue:44483", + "issue:44484", + "issue:44485", + "issue:44514", + "issue:44568", + "issue:44623", + "issue:44779", + "issue:44933", + "issue:44977", + "issue:45092", + "issue:45103", + "issue:45125", + "issue:45310", + "issue:45356", + "issue:45357", + "issue:45405", + "issue:45406", + "issue:45656", + "issue:45712", + "issue:45834" + ], + "soft_pairs": [ + "issue:30333|issue:33290", + "issue:43723|issue:44568", + "issue:43874|issue:45092", + "issue:42898|issue:45356", + "issue:44060|issue:45125", + "issue:42898|issue:44779", + "issue:45405|issue:45712", + "issue:42898|issue:44448", + "issue:44295|issue:45656", + "issue:44933|issue:45712", + "issue:44222|issue:44297", + "issue:44297|issue:45656", + "issue:43116|issue:44190", + "issue:43479|issue:43644", + "issue:44222|issue:44623", + "issue:44206|issue:45103", + "issue:42831|issue:44623", + "issue:41628|issue:42757", + "issue:22355|issue:28282", + "issue:30333|issue:36331", + "issue:44297|issue:45310", + "issue:30333|issue:34689", + "issue:43232|issue:45834", + "issue:44222|issue:44484", + "issue:43723|issue:44977", + "issue:28282|issue:30333", + "issue:44263|issue:45092", + "issue:30333|issue:35141", + "issue:44483|issue:45406", + "issue:44263|issue:44360", + "issue:44373|issue:44485", + "issue:44295|issue:44623", + "issue:44514|issue:45103", + "issue:44261|issue:45092", + "issue:45092|issue:45357" + ], + "prepared_review_unit_hash": "437a461831010131a6112a124645b25b8b32d19487fa9cb43f2a1998b9430527", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 35, + "serialized_chars": 38105, + "estimated_input_tokens": 9527, + "estimated_eval_tokens": 19310 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is very heterogeneous. The soft links mostly reflect broad topical similarity (tokenizers, save_pretrained, multimodal models, training/deepspeed, Qwen/GLM internals) rather than the same underlying bug or change, so I would not merge any of the candidate pairs.", + "confidence": 0.93, + "canonical_issue_reason": "No strong canonical duplicate hub stands out; the items span unrelated bugs. If forced to anchor the cluster, issue:44222 is the most concrete self-contained bug report, but it does not subsume the rest.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:44222 is the most actionable standalone report in the set, with a clear symptom and scope, but it is not a true umbrella for the cluster.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:30333", + "right": "issue:33290", + "accept": false, + "reason": "Both are training-related, but one is an MLflow reporting issue and the other is an OOM bug in Adafactor/DeepSpeed." + }, + { + "left": "issue:43723", + "right": "issue:44568", + "accept": false, + "reason": "Both involve tokenizer behavior in v5, but one is a load/import failure and the other is missing BOS/EOS insertion." + }, + { + "left": "issue:43874", + "right": "issue:45092", + "accept": false, + "reason": "Both touch multimodal vision paths, but one is a missing image-patch API and the other is meta-init compatibility for old checkpoints." + }, + { + "left": "issue:42898", + "right": "issue:45356", + "accept": false, + "reason": "Both are tokenizer regressions, but the affected code paths and symptoms are different." + }, + { + "left": "issue:44060", + "right": "issue:45125", + "accept": false, + "reason": "Same general Qwen/MoE area, but one is a tied-weights warning and the other is a missing tensor-parallel plan." + }, + { + "left": "issue:42898", + "right": "issue:44779", + "accept": false, + "reason": "Both concern tokenizers in v5, but they describe different models and different failure modes." + }, + { + "left": "issue:45405", + "right": "issue:45712", + "accept": false, + "reason": "Completely unrelated: a version pin issue versus leftover dummy class cleanup." + }, + { + "left": "issue:42898", + "right": "issue:44448", + "accept": false, + "reason": "Both mention v5 output changes, but one is a tokenizer-space behavior change and the other is a Pegasus model output regression." + }, + { + "left": "issue:44295", + "right": "issue:45656", + "accept": false, + "reason": "Unrelated: a buffer/position_ids bug versus a DeepSpeed double optimizer-step issue." + }, + { + "left": "issue:44933", + "right": "issue:45712", + "accept": false, + "reason": "Different areas entirely: missing image_utils import versus dummy class leakage." + }, + { + "left": "issue:44222", + "right": "issue:44297", + "accept": false, + "reason": "Both involve save_pretrained, but one is model FP8 serialization and the other is tokenizer metadata." + }, + { + "left": "issue:44297", + "right": "issue:45656", + "accept": false, + "reason": "No shared code path; tokenizer save metadata and optimizer stepping are unrelated." + }, + { + "left": "issue:43116", + "right": "issue:44190", + "accept": false, + "reason": "Both are example-script issues, but one is classification output logic and the other is local dataset loading." + }, + { + "left": "issue:43479", + "right": "issue:43644", + "accept": false, + "reason": "Different bugs: multimodal config initialization versus non-persistent buffer initialization." + }, + { + "left": "issue:44222", + "right": "issue:44623", + "accept": false, + "reason": "Both mention save_pretrained, but they concern different object types and different missing behavior." + }, + { + "left": "issue:44206", + "right": "issue:45103", + "accept": false, + "reason": "These are unrelated crashes: a feature-extractor argument mismatch versus auto-docstring annotation handling." + }, + { + "left": "issue:42831", + "right": "issue:44623", + "accept": false, + "reason": "Accuracy drift in FineGrainedFP8 is not the same problem as processor save_pretrained missing files." + }, + { + "left": "issue:41628", + "right": "issue:42757", + "accept": false, + "reason": "Both are import errors, but they come from different packages and different missing symbols." + }, + { + "left": "issue:22355", + "right": "issue:28282", + "accept": false, + "reason": "Different import failures with different root causes: missing transformers.onnx module versus missing PyTorch dependency." + }, + { + "left": "issue:30333", + "right": "issue:36331", + "accept": false, + "reason": "Both involve training, but one is MLflow reporting and the other is a Trainer API signature mismatch." + }, + { + "left": "issue:44297", + "right": "issue:45310", + "accept": false, + "reason": "Tokenizer save metadata and Qwen3.5 from_pretrained failures are different code paths and symptoms." + }, + { + "left": "issue:30333", + "right": "issue:34689", + "accept": false, + "reason": "Unrelated: MLflow/job-state reporting versus a model-loading regression for Llama 3.2 Vision." + }, + { + "left": "issue:43232", + "right": "issue:45834", + "accept": false, + "reason": "Different generation/ocr bugs with no shared underlying mechanism." + }, + { + "left": "issue:44222", + "right": "issue:44484", + "accept": false, + "reason": "Save_pretrained FP8 serialization is unrelated to max_shard_size configuration semantics." + }, + { + "left": "issue:44263", + "right": "issue:44360", + "accept": false, + "reason": "Same DSA indexer area, but the reported defects are different: torch.split handling versus missing ReLU." + }, + { + "left": "issue:44373", + "right": "issue:44485", + "accept": false, + "reason": "A docstring issue is not the same as a RoPE implementation bug." + }, + { + "left": "issue:44295", + "right": "issue:44623", + "accept": false, + "reason": "Different concerns: buffer registration/reads versus processor serialization file output." + }, + { + "left": "issue:44514", + "right": "issue:45103", + "accept": false, + "reason": "Unrelated crashes: batched chat template padding versus auto-docstring annotation introspection." + }, + { + "left": "issue:44261", + "right": "issue:45092", + "accept": false, + "reason": "Different multimodal bugs: RMS norm epsilon precision versus old checkpoint meta-init incompatibility." + }, + { + "left": "issue:44263", + "right": "issue:44360", + "accept": false, + "reason": "Related only by the DSA indexer subsystem; the code-path problem is not the same." + }, + { + "left": "issue:45092", + "right": "issue:45357", + "accept": false, + "reason": "Both relate to Qwen/vision loading, but one is meta-init compatibility and the other is save_pretrained key serialization." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 38, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:22355", + "issue:28218", + "issue:30990", + "issue:33357", + "issue:38175", + "issue:39401", + "issue:41628", + "issue:42371", + "issue:42757", + "issue:42831", + "issue:42886", + "issue:42907", + "issue:42994", + "issue:43066", + "issue:43122", + "issue:43232", + "issue:43479", + "issue:43576", + "issue:43723", + "issue:43756", + "issue:43784", + "issue:43824", + "issue:43976", + "issue:44038", + "issue:44060", + "issue:44117", + "issue:44261", + "issue:44263", + "issue:44295", + "issue:44297", + "issue:44373", + "issue:44442", + "issue:44488", + "issue:44514", + "issue:44521", + "issue:44568", + "issue:44704", + "issue:44743", + "issue:44871", + "issue:44913", + "issue:44945", + "issue:44991", + "issue:45278", + "issue:45290", + "issue:45310", + "issue:45593", + "issue:45600", + "issue:45753" + ], + "soft_pairs": [ + "issue:43723|issue:44871", + "issue:30990|issue:38175", + "issue:43723|issue:44991", + "issue:43723|issue:44442", + "issue:41628|issue:43824", + "issue:43479|issue:44913", + "issue:42831|issue:45310", + "issue:44263|issue:44373", + "issue:39401|issue:44568", + "issue:42831|issue:44038", + "issue:42757|issue:45278", + "issue:42886|issue:42994", + "issue:43232|issue:45593", + "issue:43756|issue:44060", + "issue:43576|issue:43784", + "issue:42831|issue:45753", + "issue:42886|issue:43066", + "issue:43122|issue:44521", + "issue:42371|issue:44263", + "issue:22355|issue:28218", + "issue:44117|issue:44704", + "issue:30990|issue:33357", + "issue:43232|issue:44743", + "issue:44295|issue:44297", + "issue:44704|issue:45290", + "issue:42907|issue:44945", + "issue:41628|issue:45278", + "issue:43122|issue:44514", + "issue:43976|issue:44945", + "issue:43784|issue:45600", + "issue:39401|issue:44488", + "issue:44261|issue:44373" + ], + "prepared_review_unit_hash": "30b87966ccfe3bacad7e3e73a020438f1bc6bd50a5d87403e7158e22de089491", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 32, + "serialized_chars": 36842, + "estimated_input_tokens": 9211, + "estimated_eval_tokens": 18678 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "These issues are mostly separate tokenizer, import, config, and model-behavior regressions that only overlap at a broad subsystem level. None of the soft pairs look like the same underlying bug or change, so all soft edges are rejected.", + "confidence": 0.79, + "canonical_issue_reason": "issue:43723 is the broadest and most central tokenizer-loading regression in this set, so it is the best umbrella issue even though the others are distinct bugs.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43723 has the clearest general scope around AutoTokenizer loading failures in v5 and the strongest fit as a representative issue for the cluster.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43723", + "right": "issue:44871", + "accept": false, + "reason": "Both are tokenizer-related, but one is a general loading regression and the other is an eos_token_id configuration mismatch; not the same bug." + }, + { + "left": "issue:30990", + "right": "issue:38175", + "accept": false, + "reason": "Unrelated symptoms: Sentence Transformers loading hang versus zero probabilities in a SigLIP model." + }, + { + "left": "issue:43723", + "right": "issue:44991", + "accept": false, + "reason": "Both mention tokenizer loading, but they concern different models and likely different failure modes." + }, + { + "left": "issue:43723", + "right": "issue:44442", + "accept": false, + "reason": "General AutoTokenizer load issue versus a specific FastSpeech2ConformerTokenizer failure; not the same underlying defect." + }, + { + "left": "issue:41628", + "right": "issue:43824", + "accept": false, + "reason": "Both are import errors, but for different symbols and model classes; too broad to treat as duplicates." + }, + { + "left": "issue:43479", + "right": "issue:44913", + "accept": false, + "reason": "Different config bugs: default sub-config initialization versus rotary_pct not persisting on reload." + }, + { + "left": "issue:42831", + "right": "issue:45310", + "accept": false, + "reason": "Accuracy regression in FineGrainedFP8 and a from_pretrained error for Qwen3.5 Moe are unrelated." + }, + { + "left": "issue:44263", + "right": "issue:44373", + "accept": false, + "reason": "A tensor-splitting bug and a wrong docstring are not the same issue." + }, + { + "left": "issue:39401", + "right": "issue:44568", + "accept": false, + "reason": "Both involve tokenization behavior, but one is offset mapping and the other is add_special_tokens/BOS-EOS handling." + }, + { + "left": "issue:42831", + "right": "issue:44038", + "accept": false, + "reason": "FineGrainedFP8 accuracy issue and Qwen3-VL-Moe loading/bug report are different problems." + }, + { + "left": "issue:42757", + "right": "issue:45278", + "accept": false, + "reason": "A missing import from huggingface_hub and a broad set of import errors are only superficially related." + }, + { + "left": "issue:42886", + "right": "issue:42994", + "accept": false, + "reason": "Offline tokenizer cache loading and quantized model saving are distinct code paths." + }, + { + "left": "issue:43232", + "right": "issue:45593", + "accept": false, + "reason": "Generation kwargs handling and D-FINE auxiliary-loss behavior are unrelated." + }, + { + "left": "issue:43756", + "right": "issue:44060", + "accept": false, + "reason": "Different Qwen/Smollm model bugs with no clear shared code path." + }, + { + "left": "issue:43576", + "right": "issue:43784", + "accept": false, + "reason": "A broken CLI command and a NameError during sentence-transformers import are unrelated." + }, + { + "left": "issue:42831", + "right": "issue:45753", + "accept": false, + "reason": "FP8 accuracy regression and an infinite loop on a specific image are different issues." + }, + { + "left": "issue:42886", + "right": "issue:43066", + "accept": false, + "reason": "Offline cache loading and wrong tokenizer decoder type are tokenizer-related but not the same bug." + }, + { + "left": "issue:43122", + "right": "issue:44521", + "accept": false, + "reason": "Both touch tokenization, but one is a version-to-version tokenization change and the other is assistant_mask generation for multimodal chat templates." + }, + { + "left": "issue:42371", + "right": "issue:44263", + "accept": false, + "reason": "TF32 settings guidance and a torch.split return-value bug are unrelated." + }, + { + "left": "issue:22355", + "right": "issue:28218", + "accept": false, + "reason": "A missing transformers.onnx module and an added-token spacing bug are unrelated." + }, + { + "left": "issue:44117", + "right": "issue:44704", + "accept": false, + "reason": "Tokenizer mapping handling and AutoProcessor kwargs forwarding are different loader bugs." + }, + { + "left": "issue:30990", + "right": "issue:33357", + "accept": false, + "reason": "A loading hang and a MacOS bus error are distinct failures." + }, + { + "left": "issue:43232", + "right": "issue:44743", + "accept": false, + "reason": "Generation state update after sync_gpus and recurrent-state resetting in qwen3 modular code are not the same defect." + }, + { + "left": "issue:44295", + "right": "issue:44297", + "accept": false, + "reason": "Reading position_ids as a buffer and tokenizer_class mismatch on save are unrelated tokenizer/model issues." + }, + { + "left": "issue:44704", + "right": "issue:45290", + "accept": false, + "reason": "AutoProcessor kwargs forwarding and apply_chat_template crashing on tool-call messages are different code paths." + }, + { + "left": "issue:42907", + "right": "issue:44945", + "accept": false, + "reason": "Quantized model saving failure and incorrect output under pipeline parallelism are unrelated." + }, + { + "left": "issue:41628", + "right": "issue:45278", + "accept": false, + "reason": "A specific AutoImageProcessor import error and broad post-upgrade import errors are not the same bug." + }, + { + "left": "issue:43122", + "right": "issue:44514", + "accept": false, + "reason": "Version-dependent tokenization differences and batched apply_chat_template crashes are different issues." + }, + { + "left": "issue:43976", + "right": "issue:44945", + "accept": false, + "reason": "Python version compatibility and pipeline-parallelism output correctness are unrelated." + }, + { + "left": "issue:43784", + "right": "issue:45600", + "accept": false, + "reason": "Import-time NameError in sentence-transformers and removed auto_mappings references are different problems." + }, + { + "left": "issue:39401", + "right": "issue:44488", + "accept": false, + "reason": "A Qwen3 offset-mapping bug and a failure to load a specific model are not the same issue." + }, + { + "left": "issue:44261", + "right": "issue:44373", + "accept": false, + "reason": "A precision bug in MLA q_a_layernorm and a wrong docstring for position_ids are unrelated." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 39, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:22355", + "issue:29127", + "issue:30990", + "issue:36296", + "issue:38617", + "issue:39401", + "issue:42371", + "issue:42673", + "issue:42757", + "issue:42907", + "issue:43064", + "issue:43066", + "issue:43116", + "issue:43122", + "issue:43278", + "issue:43408", + "issue:43452", + "issue:43475", + "issue:43493", + "issue:43673", + "issue:43784", + "issue:43824", + "issue:43874", + "issue:43937", + "issue:43976", + "issue:43992", + "issue:44038", + "issue:44222", + "issue:44295", + "issue:44297", + "issue:44351", + "issue:44360", + "issue:44485", + "issue:44521", + "issue:44625", + "issue:44704", + "issue:44779", + "issue:44829", + "issue:44945", + "issue:44964", + "issue:44977", + "issue:45072", + "issue:45200", + "issue:45216", + "issue:45278", + "issue:45446", + "issue:45478", + "issue:45753" + ], + "soft_pairs": [ + "issue:43408|issue:43475", + "issue:42907|issue:43937", + "issue:44360|issue:44485", + "issue:44625|issue:44829", + "issue:39401|issue:44521", + "issue:43116|issue:45200", + "issue:29127|issue:30990", + "issue:44038|issue:44297", + "issue:42907|issue:44964", + "issue:43122|issue:44779", + "issue:44038|issue:44295", + "issue:36296|issue:43116", + "issue:42907|issue:43452", + "issue:43824|issue:44351", + "issue:44038|issue:44222", + "issue:45216|issue:45478", + "issue:42757|issue:44351", + "issue:43493|issue:45753", + "issue:43278|issue:44945", + "issue:42907|issue:44222", + "issue:43122|issue:44977", + "issue:43992|issue:44704", + "issue:43784|issue:45446", + "issue:43066|issue:44779", + "issue:38617|issue:45278", + "issue:42673|issue:45072", + "issue:43937|issue:44485", + "issue:22355|issue:29127", + "issue:42371|issue:43874", + "issue:43064|issue:43673", + "issue:43976|issue:45446", + "issue:43116|issue:43493" + ], + "prepared_review_unit_hash": "57afcd8047dc4ffa57986e6bc57645097e1888a68b534d60563f8a93fe19c1f8", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 32, + "serialized_chars": 36958, + "estimated_input_tokens": 9240, + "estimated_eval_tokens": 18736 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly heterogeneous: the soft pairs share broad themes like imports, tokenization, or Qwen/SAM/GLM models, but they do not appear to be the same concrete bug or change. I would keep them separate; no PRs are present.", + "confidence": 0.71, + "canonical_issue_reason": "issue:45278 is the broadest umbrella for the import-regression subset (many import errors after updating to 5.5.0), making it the best representative issue in this mixed cluster.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45278 best captures the most central recurring theme here: upgrade-related import failures across multiple reports. It is still not a duplicate of the others, just the most useful anchor issue.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43408", + "right": "issue:43475", + "accept": false, + "reason": "Same model family, but different bugs: a model-type warning vs a missing attribute on Sam3VisionEncoderOutput." + }, + { + "left": "issue:42907", + "right": "issue:43937", + "accept": false, + "reason": "Different failures: dequantized save_pretrained vs GLM-5 GenerationConfig validation." + }, + { + "left": "issue:44360", + "right": "issue:44485", + "accept": false, + "reason": "Unrelated: DSA indexer ReLU discussion vs GLM-5 RoPE implementation." + }, + { + "left": "issue:44625", + "right": "issue:44829", + "accept": false, + "reason": "Different code paths: Qwen3.5 num_labels propagation vs flash_attention_3 degenerate training." + }, + { + "left": "issue:39401", + "right": "issue:44521", + "accept": false, + "reason": "Tokenizer offset_mapping bug vs multimodal apply_chat_template assistant mask bug." + }, + { + "left": "issue:43116", + "right": "issue:45200", + "accept": false, + "reason": "Example-script multi-label output bug vs Gemma 4 mm_token_type_ids defaulting issue." + }, + { + "left": "issue:29127", + "right": "issue:30990", + "accept": false, + "reason": "LayoutLMv3 error-message clarity vs Sentence Transformers loading hang; not the same defect." + }, + { + "left": "issue:44038", + "right": "issue:44297", + "accept": false, + "reason": "Qwen3-VL-Moe regression vs tokenizer save_pretrained metadata mismatch." + }, + { + "left": "issue:42907", + "right": "issue:44964", + "accept": false, + "reason": "Different model-loading/saving issues: dequantized save failure vs Phi-4 multimodal load failure." + }, + { + "left": "issue:43122", + "right": "issue:44779", + "accept": false, + "reason": "Both are tokenizer regressions, but they concern different models and likely different root causes." + }, + { + "left": "issue:44038", + "right": "issue:44295", + "accept": false, + "reason": "Different bugs: Qwen3-VL-Moe regression vs position_ids buffer access." + }, + { + "left": "issue:36296", + "right": "issue:43116", + "accept": false, + "reason": "Tensor-parallel training bug vs classification example returning empty results." + }, + { + "left": "issue:42907", + "right": "issue:43452", + "accept": false, + "reason": "Different loading/saving paths: dequantized save_pretrained vs gguf_file loading failure." + }, + { + "left": "issue:43824", + "right": "issue:44351", + "accept": false, + "reason": "Both import errors, but for different symbols and causes; not the same underlying bug." + }, + { + "left": "issue:44038", + "right": "issue:44222", + "accept": false, + "reason": "Same broad MoE area, but Qwen3-VL-Moe regression and FP8 save_pretrained are different issues." + }, + { + "left": "issue:45216", + "right": "issue:45478", + "accept": false, + "reason": "Both involve Qwen3.5 regressions, but one is save_pretrained checkpoint corruption and the other is from_pretrained failure." + }, + { + "left": "issue:42757", + "right": "issue:44351", + "accept": false, + "reason": "Distinct import failures from different packages/symbols, not a shared root cause." + }, + { + "left": "issue:43493", + "right": "issue:45753", + "accept": false, + "reason": "Different model families and symptoms: SigLIP2 discrepancy vs Qwen3_5 image-specific infinite loop." + }, + { + "left": "issue:43278", + "right": "issue:44945", + "accept": false, + "reason": "Dtype mismatch between train/eval vs incorrect output under pipeline parallelism." + }, + { + "left": "issue:42907", + "right": "issue:44222", + "accept": false, + "reason": "Both mention save_pretrained, but dequantized save and FP8 MoE save are different bugs." + }, + { + "left": "issue:43122", + "right": "issue:44977", + "accept": false, + "reason": "Tokenizer behavior regression vs generation failure with flash-attention; different concrete problems." + }, + { + "left": "issue:43992", + "right": "issue:44704", + "accept": false, + "reason": "Separate loading bugs: missing embed_tokens.weight vs cached_file kwargs not forwarded." + }, + { + "left": "issue:43784", + "right": "issue:45446", + "accept": false, + "reason": "Import-related, but one is a NameError in sentence-transformers and the other is a version check bug in flex_attention." + }, + { + "left": "issue:43066", + "right": "issue:44779", + "accept": false, + "reason": "Different tokenizer regressions with different symptoms and likely different fixes." + }, + { + "left": "issue:38617", + "right": "issue:45278", + "accept": false, + "reason": "45278 is a broad import-regression report, but 38617 is a specific missing symbol; not the same concrete bug." + }, + { + "left": "issue:42673", + "right": "issue:45072", + "accept": false, + "reason": "VRAM leak in dataloader threads vs bfloat16 dtype mismatch in inference." + }, + { + "left": "issue:43937", + "right": "issue:44485", + "accept": false, + "reason": "GenerationConfig validation issue vs RoPE implementation discussion." + }, + { + "left": "issue:22355", + "right": "issue:29127", + "accept": false, + "reason": "No module named transformers.onnx vs LayoutLMv3 box-information error message; unrelated." + }, + { + "left": "issue:42371", + "right": "issue:43874", + "accept": false, + "reason": "TF32 API/settings issue vs missing image-patch count method in GLM46VImageProcessorFast." + }, + { + "left": "issue:43064", + "right": "issue:43673", + "accept": false, + "reason": "Different training/runtime failures: FSDP2 optimizer-state bug vs missing GenerationMixin cache in chunked prefill." + }, + { + "left": "issue:43976", + "right": "issue:45446", + "accept": false, + "reason": "Python-version compatibility issue vs incorrect PyTorch version check for AuxRequest import." + }, + { + "left": "issue:43116", + "right": "issue:43493", + "accept": false, + "reason": "Example-script classification bug vs SigLIP2 implementation discrepancy." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 40, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:33453", + "issue:38617", + "issue:39401", + "issue:41628", + "issue:42222", + "issue:42371", + "issue:42490", + "issue:42617", + "issue:42673", + "issue:43116", + "issue:43408", + "issue:43421", + "issue:43441", + "issue:43450", + "issue:43582", + "issue:43742", + "issue:43825", + "issue:43827", + "issue:43906", + "issue:43937", + "issue:43986", + "issue:44038", + "issue:44075", + "issue:44117", + "issue:44261", + "issue:44291", + "issue:44410", + "issue:44484", + "issue:44779", + "issue:44829", + "issue:44933", + "issue:44936", + "issue:44977", + "issue:45071", + "issue:45081", + "issue:45083", + "issue:45216", + "issue:45278", + "issue:45310", + "issue:45341", + "issue:45357", + "issue:45405", + "issue:45446", + "issue:45478", + "issue:45542", + "issue:45561", + "issue:45563", + "issue:45588" + ], + "soft_pairs": [ + "issue:42222|issue:43441", + "issue:44075|issue:44936", + "issue:43116|issue:43825", + "issue:45357|issue:45478", + "issue:44038|issue:45478", + "issue:45083|issue:45446", + "issue:43937|issue:45563", + "issue:44117|issue:45588", + "issue:33453|issue:39401", + "issue:43827|issue:45446", + "issue:39401|issue:44779", + "issue:43937|issue:44977", + "issue:42673|issue:43906", + "issue:44291|issue:45071", + "issue:43450|issue:43986", + "issue:45405|issue:45446", + "issue:45341|issue:45561", + "issue:44117|issue:45081", + "issue:38617|issue:41628", + "issue:39401|issue:43421", + "issue:44410|issue:45310", + "issue:44484|issue:45310", + "issue:44829|issue:45216", + "issue:42371|issue:44261", + "issue:42673|issue:43582", + "issue:43408|issue:43742", + "issue:42617|issue:45341", + "issue:44933|issue:45278", + "issue:42222|issue:45542", + "issue:42490|issue:42673" + ], + "prepared_review_unit_hash": "5d0bdbcdb34ece144caccd9921d5e6a848e590a8662f160eaa2d39da66a95cb7", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 30, + "serialized_chars": 36465, + "estimated_input_tokens": 9117, + "estimated_eval_tokens": 18490 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "These are mostly separate regression reports across different models, tokenizer paths, and runtime subsystems. I do not see any soft pair that clearly describes the same underlying bug, so I would not merge any of them as duplicates.", + "confidence": 0.39, + "canonical_issue_reason": "issue:45278 is the broadest umbrella report in the set: a current open issue about widespread import failures after upgrading to v5.5.0, which best matches the cluster\u2019s import-regression theme.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45278 is the most suitable representative issue because it is open, broad, and could subsume related import-error regressions better than the highly specific model- or API-specific reports.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:42222", + "right": "issue:43441", + "accept": false, + "reason": "Different failures: VitPose model loading breakage vs Ministral-3 FlashAttention incompatibility." + }, + { + "left": "issue:44075", + "right": "issue:44936", + "accept": false, + "reason": "Unrelated training bugs: SGD args ignored vs trainer.evaluate() failing after train()." + }, + { + "left": "issue:43116", + "right": "issue:43825", + "accept": false, + "reason": "Different concerns: multi-label example returning empty results vs pipeline() error text about unsupported translation tasks." + }, + { + "left": "issue:45357", + "right": "issue:45478", + "accept": false, + "reason": "Same model family, but different code paths: save_pretrained key layout vs from_pretrained loading error." + }, + { + "left": "issue:44038", + "right": "issue:45478", + "accept": false, + "reason": "Both mention Qwen/MoE, but the model/version and failure mode differ; not the same concrete bug." + }, + { + "left": "issue:45083", + "right": "issue:45446", + "accept": false, + "reason": "Completely different subsystems: qwen3_omni_moe helper logic vs flex_attention import version check." + }, + { + "left": "issue:43937", + "right": "issue:45563", + "accept": false, + "reason": "GenerationConfig validation error is unrelated to the stale warning for num_return_sequences." + }, + { + "left": "issue:44117", + "right": "issue:45588", + "accept": false, + "reason": "Tokenizer mapping None bug vs flash_attention crash on s_aux=None are unrelated." + }, + { + "left": "issue:33453", + "right": "issue:39401", + "accept": false, + "reason": "Both tokenizer-related, but one is loading regression and the other is offset_mapping correctness for Qwen3." + }, + { + "left": "issue:43827", + "right": "issue:45446", + "accept": false, + "reason": "Docs referencing pipeline() vs a PyTorch version check bug are unrelated." + }, + { + "left": "issue:39401", + "right": "issue:44779", + "accept": false, + "reason": "Different tokenizer regressions affecting different models and symptoms." + }, + { + "left": "issue:43937", + "right": "issue:44977", + "accept": false, + "reason": "Invalid GenerationConfig vs flash-attention generation problems are different bugs." + }, + { + "left": "issue:42673", + "right": "issue:43906", + "accept": false, + "reason": "VRAM leak in dataloader threads vs an isolated reproduction issue; no shared concrete defect." + }, + { + "left": "issue:44291", + "right": "issue:45071", + "accept": false, + "reason": "init_empty_weights argument handling vs PretrainedConfig type checking are separate code-path regressions." + }, + { + "left": "issue:43450", + "right": "issue:43986", + "accept": false, + "reason": "Video processor batch-shape bug vs missing torchvision crash when using AutoProcessor are different failures." + }, + { + "left": "issue:45405", + "right": "issue:45446", + "accept": false, + "reason": "PEFT version bump issue and flex_attention import check are unrelated." + }, + { + "left": "issue:45341", + "right": "issue:45561", + "accept": false, + "reason": "A testing_utils bug and a pytest-xdist race in patched testing utils are not the same issue." + }, + { + "left": "issue:44117", + "right": "issue:45081", + "accept": false, + "reason": "Tokenizer mapping fallback bug vs Mistral regex patch crash are different tokenizer paths." + }, + { + "left": "issue:38617", + "right": "issue:41628", + "accept": false, + "reason": "Both are import errors, but they involve different symbols and different underlying regressions." + }, + { + "left": "issue:39401", + "right": "issue:43421", + "accept": false, + "reason": "Qwen3 offset_mapping bug vs runtime special-token/post-processor update issue are not the same defect." + }, + { + "left": "issue:44410", + "right": "issue:45310", + "accept": false, + "reason": "Missing projections in qwen3next is unrelated to Qwen3.5 from_pretrained loading failure." + }, + { + "left": "issue:44484", + "right": "issue:45310", + "accept": false, + "reason": "save_pretrained shard-size question vs from_pretrained error are not duplicates." + }, + { + "left": "issue:44829", + "right": "issue:45216", + "accept": false, + "reason": "FlashAttention3 training degeneration is unrelated to incorrect checkpoint saving." + }, + { + "left": "issue:42371", + "right": "issue:44261", + "accept": false, + "reason": "TF32 API guidance issue vs MLA rms_norm_eps precision bug are different topics." + }, + { + "left": "issue:42673", + "right": "issue:43582", + "accept": false, + "reason": "VRAM leak in threads vs Apple Silicon TypeError in caching_allocator_warmup are unrelated." + }, + { + "left": "issue:43408", + "right": "issue:43742", + "accept": false, + "reason": "Model-type warning for sam3 vs key error loading MobileLLM are unrelated." + }, + { + "left": "issue:42617", + "right": "issue:45341", + "accept": false, + "reason": "3d_parallel.py execution failure and a testing_utils bug do not share the same code path." + }, + { + "left": "issue:44933", + "right": "issue:45278", + "accept": false, + "reason": "Related only by broad import-error theme; 45278 is an umbrella report, not the same concrete bug as the image_utils import issue." + }, + { + "left": "issue:42222", + "right": "issue:45542", + "accept": false, + "reason": "VitPose breakage and a TensorFlow backend error from tensorboard-only installs are unrelated." + }, + { + "left": "issue:42490", + "right": "issue:42673", + "accept": false, + "reason": "Trailing-slash local model path warning vs VRAM leak in dataloader threads are different bugs." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 41, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:38617", + "issue:41628", + "issue:42175", + "issue:42222", + "issue:42371", + "issue:42757", + "issue:42831", + "issue:43012", + "issue:43066", + "issue:43097", + "issue:43122", + "issue:43335", + "issue:43408", + "issue:43441", + "issue:43452", + "issue:43531", + "issue:43577", + "issue:43673", + "issue:43874", + "issue:43931", + "issue:43937", + "issue:43976", + "issue:43986", + "issue:44038", + "issue:44075", + "issue:44117", + "issue:44220", + "issue:44261", + "issue:44263", + "issue:44368", + "issue:44410", + "issue:44483", + "issue:44484", + "issue:44485", + "issue:44493", + "issue:44496", + "issue:44829", + "issue:44987", + "issue:45003", + "issue:45071", + "issue:45245", + "issue:45341", + "issue:45372", + "issue:45459", + "issue:45464", + "issue:45478", + "issue:45542", + "issue:45656" + ], + "soft_pairs": [ + "issue:43937|issue:44496", + "issue:42175|issue:42222", + "issue:44483|issue:45464", + "issue:38617|issue:44220", + "issue:42371|issue:42831", + "issue:43066|issue:43122", + "issue:44075|issue:45656", + "issue:43673|issue:45341", + "issue:43012|issue:43577", + "issue:43874|issue:44261", + "issue:44484|issue:45478", + "issue:44410|issue:45478", + "issue:43097|issue:44493", + "issue:44263|issue:44485", + "issue:41628|issue:45071", + "issue:43012|issue:43408", + "issue:42222|issue:45459", + "issue:43441|issue:44829", + "issue:45071|issue:45542", + "issue:43931|issue:44410", + "issue:43335|issue:45071", + "issue:43931|issue:44038", + "issue:45003|issue:45341", + "issue:43097|issue:43452", + "issue:43986|issue:45341", + "issue:44117|issue:44987", + "issue:43986|issue:44484", + "issue:43408|issue:44368", + "issue:43874|issue:44263", + "issue:43976|issue:45071", + "issue:43986|issue:45372", + "issue:44829|issue:45245", + "issue:43531|issue:45341", + "issue:38617|issue:42757", + "issue:44261|issue:44485" + ], + "prepared_review_unit_hash": "51fbfce399d3084a546659130372e6d2a6c5b5020106574d5280cd7334dc9428", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 35, + "serialized_chars": 37997, + "estimated_input_tokens": 9500, + "estimated_eval_tokens": 19256 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is a loose mix of Transformers regressions and import/model-loading bugs that only look similar at the title level. The soft edges mostly share generic v5 migration or backend wording, but they describe different code paths and failure modes, so none should be merged as duplicates.", + "confidence": 0.64, + "canonical_issue_reason": "If a single anchor is needed, issue 45372 is the most active and current report, but the cluster is too heterogeneous to have a true canonical duplicate target.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 45372 is the best single issue to represent the cluster only because it is open, concrete, and has follow-up traffic; it is not a strong duplicate hub for the other reports.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43937", + "right": "issue:44496", + "accept": false, + "reason": "Both are v5-era validation/loading complaints, but one is a GenerationConfig error for GLM-5 and the other is an unrecognized model/config problem for OLMo; different failures." + }, + { + "left": "issue:42175", + "right": "issue:42222", + "accept": false, + "reason": "These are unrelated backend/model issues: TensorFlow not being included with a torch install versus broken vitpose model files." + }, + { + "left": "issue:44483", + "right": "issue:45464", + "accept": false, + "reason": "Both mention chat/completions, but one is a general API request rejection and the other is a streaming inference failure on a specific Qwen model." + }, + { + "left": "issue:38617", + "right": "issue:44220", + "accept": false, + "reason": "These are different import/audio bugs: a missing transformers symbol versus a feature extraction helper issue." + }, + { + "left": "issue:42371", + "right": "issue:42831", + "accept": false, + "reason": "TF32 configuration guidance and FineGrainedFP8 accuracy are different numeric-behavior problems, not the same bug." + }, + { + "left": "issue:43066", + "right": "issue:43122", + "accept": false, + "reason": "Related to tokenizer behavior in v5, but one is a wrong decoder type and the other is a tokenization change regression; not the same underlying defect." + }, + { + "left": "issue:44075", + "right": "issue:45656", + "accept": false, + "reason": "Both involve optimizers stepping incorrectly, but the concrete issue differs: unused SGD args versus deepspeed calling step twice." + }, + { + "left": "issue:43673", + "right": "issue:45341", + "accept": false, + "reason": "Generation cache availability in v5 is unrelated to a small testing_utils bug." + }, + { + "left": "issue:43012", + "right": "issue:43577", + "accept": false, + "reason": "Both concern precision/dtype handling, but one is a PyTorch warning during bf16 compilation and the other is BLIP2 loading leaving submodules in float32." + }, + { + "left": "issue:43874", + "right": "issue:44261", + "accept": false, + "reason": "Both touch numerical precision, but one is a missing image-patch helper in GLM46V and the other is an MLA layernorm epsilon mismatch." + }, + { + "left": "issue:44484", + "right": "issue:45478", + "accept": false, + "reason": "Both are Qwen3.5-related failures, but one asks about save_pretrained shard sizing while the other is a from_pretrained error in a MoE model." + }, + { + "left": "issue:44410", + "right": "issue:45478", + "accept": false, + "reason": "Layer-0 missing projections in qwen3next is a model-architecture bug, not the same as Qwen3.5 MoE loading failure." + }, + { + "left": "issue:43097", + "right": "issue:44493", + "accept": false, + "reason": "These are both v5 migration regressions, but one is removal of an encoder-decoder tie flag and the other is unexpected position-id keys during loading." + }, + { + "left": "issue:44263", + "right": "issue:44485", + "accept": false, + "reason": "Different GLM-family bugs: torch.split return handling versus RoPE implementation concerns." + }, + { + "left": "issue:41628", + "right": "issue:45071", + "accept": false, + "reason": "Both are API/regression reports, but one is a missing AutoImageProcessor import and the other is a PreTrainedConfig type-checking break." + }, + { + "left": "issue:43012", + "right": "issue:43408", + "accept": false, + "reason": "Both mention warnings, but one is bf16 compile precision and the other is a model-type mismatch warning for sam3_video/sam3_tracker." + }, + { + "left": "issue:42222", + "right": "issue:45459", + "accept": false, + "reason": "A broken vitpose model and a tokenizer error-hiding import_protobuf_decode_error problem are unrelated." + }, + { + "left": "issue:43441", + "right": "issue:44829", + "accept": false, + "reason": "Ministral FlashAttention failure and flash_attention_3 training degeneration are both attention-related, but they are different model/code paths and symptoms." + }, + { + "left": "issue:45071", + "right": "issue:45542", + "accept": false, + "reason": "PreTrainedConfig type checking and a missing TensorFlow backend caused by tensorboard-only installs are not the same regression." + }, + { + "left": "issue:43931", + "right": "issue:44410", + "accept": false, + "reason": "These are separate model-loading shape mismatch and qwen3next projection bugs." + }, + { + "left": "issue:43335", + "right": "issue:45071", + "accept": false, + "reason": "SwitchTransformers sparse-layer construction and PreTrainedConfig type checking are distinct config/constructor issues." + }, + { + "left": "issue:43931", + "right": "issue:44038", + "accept": false, + "reason": "Qwen3-VL weight-shape mismatch and Qwen3-VL-Moe from_pretrained failures are different model variants and failure modes." + }, + { + "left": "issue:45003", + "right": "issue:45341", + "accept": false, + "reason": "Unsafe sys.modules access in modeling_utils is unrelated to the minor testing_utils bug." + }, + { + "left": "issue:43097", + "right": "issue:43452", + "accept": false, + "reason": "A removed tie_embeddings flag and gguf_file breaking tokenizer/model loading are separate regression areas." + }, + { + "left": "issue:43986", + "right": "issue:45341", + "accept": false, + "reason": "A missing torchvision dependency path in AutoProcessor is unrelated to a small testing utility bug." + }, + { + "left": "issue:44117", + "right": "issue:44987", + "accept": false, + "reason": "TOKENIZER_MAPPING_NAMES returning None and loading physical-intelligence/fast are different tokenizer versus model-loading failures." + }, + { + "left": "issue:43986", + "right": "issue:44484", + "accept": false, + "reason": "AutoProcessor crashing without torchvision and max_shard_size behavior are unrelated." + }, + { + "left": "issue:43408", + "right": "issue:44368", + "accept": false, + "reason": "Model-type mismatch warnings and tie_word_embeddings warnings are separate configuration warnings." + }, + { + "left": "issue:43874", + "right": "issue:44263", + "accept": false, + "reason": "Both are GLM-related, but one is a missing image-patch helper and the other is a torch.split return-value issue." + }, + { + "left": "issue:43976", + "right": "issue:45071", + "accept": false, + "reason": "Python version support in 5.1.0 and PreTrainedConfig type checking are unrelated release regressions." + }, + { + "left": "issue:43986", + "right": "issue:45372", + "accept": false, + "reason": "AutoProcessor without torchvision and the ReasoningEffort import error affect different processor/dependency paths." + }, + { + "left": "issue:44829", + "right": "issue:45245", + "accept": false, + "reason": "FlashAttention-3 training degeneration and the 2^24 category limit are completely different problems." + }, + { + "left": "issue:43531", + "right": "issue:45341", + "accept": false, + "reason": "Qwen3-MoE sliding_window behavior is unrelated to the testing_utils bug." + }, + { + "left": "issue:38617", + "right": "issue:42757", + "accept": false, + "reason": "Both are import errors, but they involve different missing symbols from different packages and code paths." + }, + { + "left": "issue:44261", + "right": "issue:44485", + "accept": false, + "reason": "MLA q_a_layernorm epsilon precision and GLM-5 RoPE implementation are separate numerical/model-architecture issues." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 42, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:36683", + "issue:38617", + "issue:42490", + "issue:42757", + "issue:42886", + "issue:42907", + "issue:43012", + "issue:43064", + "issue:43066", + "issue:43097", + "issue:43208", + "issue:43352", + "issue:43381", + "issue:43502", + "issue:43526", + "issue:43550", + "issue:43582", + "issue:43618", + "issue:43761", + "issue:43824", + "issue:43825", + "issue:43873", + "issue:43901", + "issue:43906", + "issue:43976", + "issue:43986", + "issue:43992", + "issue:44038", + "issue:44062", + "issue:44117", + "issue:44222", + "issue:44315", + "issue:44351", + "issue:44368", + "issue:44380", + "issue:44410", + "issue:44483", + "issue:44484", + "issue:44514", + "issue:44655", + "issue:44683", + "issue:44704", + "issue:45103", + "issue:45375", + "issue:45478", + "issue:45479", + "issue:45588", + "issue:45753" + ], + "soft_pairs": [ + "issue:43064|issue:44315", + "issue:38617|issue:44351", + "issue:44410|issue:45375", + "issue:44062|issue:44117", + "issue:43097|issue:45588", + "issue:44704|issue:45103", + "issue:44038|issue:44410", + "issue:43502|issue:43992", + "issue:43986|issue:44117", + "issue:43097|issue:43873", + "issue:43824|issue:43986", + "issue:43066|issue:43906", + "issue:43097|issue:43901", + "issue:43825|issue:44655", + "issue:43992|issue:44484", + "issue:43208|issue:43550", + "issue:43012|issue:43550", + "issue:43352|issue:44380", + "issue:43012|issue:44368", + "issue:42886|issue:43381", + "issue:42907|issue:45753", + "issue:42907|issue:44655", + "issue:44380|issue:45479", + "issue:43352|issue:43976", + "issue:43064|issue:44380", + "issue:43208|issue:43526", + "issue:44483|issue:44514", + "issue:36683|issue:43582", + "issue:44222|issue:45478", + "issue:42490|issue:43012", + "issue:43618|issue:43761", + "issue:43012|issue:44683", + "issue:42757|issue:43824", + "issue:43618|issue:44380", + "issue:38617|issue:43502" + ], + "prepared_review_unit_hash": "aa33f3466cdd8039ab38eb536d8dd3e1f429f0101b9aa9f0dad44c6c2ff10797", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 35, + "serialized_chars": 38058, + "estimated_input_tokens": 9515, + "estimated_eval_tokens": 19286 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This is a loose similarity cluster of mostly unrelated Transformers issues: import errors, model-specific regressions, tokenizer/pipeline bugs, and training/runtime edge cases. I do not see any soft pair that is clearly the same concrete bug/change, so all soft edges should be rejected.", + "confidence": 0.91, + "canonical_issue_reason": "No single canonical issue fits the whole cluster; the items span unrelated bugs and regressions rather than one shared defect.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43873 is the most active open item in the set and has broad user impact around quantized/offloaded loading, so it is the best representative anchor among these heterogeneous issues.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43064", + "right": "issue:44315", + "accept": false, + "reason": "Different problems: distributed Trainer/FSDP2 optimizer-state corruption vs Liger Kernel application with model_init." + }, + { + "left": "issue:38617", + "right": "issue:44351", + "accept": false, + "reason": "Both are import errors, but for different missing symbols and different code paths." + }, + { + "left": "issue:44410", + "right": "issue:45375", + "accept": false, + "reason": "Different model configs/fields: missing qwen3next projections vs missing deepstack_visual_indexes in Qwen3_5MoeVisionConfig." + }, + { + "left": "issue:44062", + "right": "issue:44117", + "accept": false, + "reason": "Tokenizer-related, but one is AddedToken argument duplication and the other is TOKENIZER_MAPPING_NAMES returning None." + }, + { + "left": "issue:43097", + "right": "issue:45588", + "accept": false, + "reason": "Unrelated regressions: removed tie_embeddings_and_encoder_decoder vs flash_attention.py crashing on s_aux=None." + }, + { + "left": "issue:44704", + "right": "issue:45103", + "accept": false, + "reason": "Different utilities and failures: cached_file kwargs forwarding vs auto_docstring annotation handling." + }, + { + "left": "issue:44038", + "right": "issue:44410", + "accept": false, + "reason": "Different Qwen model families and different failures; not the same underlying bug." + }, + { + "left": "issue:43502", + "right": "issue:43992", + "accept": false, + "reason": "One is offline/local_files_only network leakage; the other is a missing embed_tokens.weight on load." + }, + { + "left": "issue:43986", + "right": "issue:44117", + "accept": false, + "reason": "Different areas: AutoProcessor/video-model dependency handling vs tokenizer mapping None handling." + }, + { + "left": "issue:43097", + "right": "issue:43873", + "accept": false, + "reason": "Different issues: API removal warning vs quantization/offloading behavior." + }, + { + "left": "issue:43824", + "right": "issue:43986", + "accept": false, + "reason": "Both touch VL/processor loading, but one is an import failure and the other is a missing torchvision dependency crash." + }, + { + "left": "issue:43066", + "right": "issue:43906", + "accept": false, + "reason": "Wrong tokenizer decoder type in v5 is unrelated to the isolated reproduction issue." + }, + { + "left": "issue:43097", + "right": "issue:43901", + "accept": false, + "reason": "API removal vs documentation/behavior mismatch for return_all_scores; not the same defect." + }, + { + "left": "issue:43825", + "right": "issue:44655", + "accept": false, + "reason": "Different pipeline failures: misleading task support message vs inability to save Pipeline objects." + }, + { + "left": "issue:43992", + "right": "issue:44484", + "accept": false, + "reason": "Missing model weight loading vs a max_shard_size question; no shared underlying bug." + }, + { + "left": "issue:43208", + "right": "issue:43550", + "accept": false, + "reason": "Different model/runtime bugs: xLSTM training blockers vs Bamba compile+SDPA failure." + }, + { + "left": "issue:43012", + "right": "issue:43550", + "accept": false, + "reason": "A compile warning on bfloat16 models is not the same as a compile failure on Bamba with SDPA." + }, + { + "left": "issue:43352", + "right": "issue:44380", + "accept": false, + "reason": "Different attention-related bugs in different models: unsupported Flash Attention 2.0 vs GPT2 attention scaling being ignored." + }, + { + "left": "issue:43012", + "right": "issue:44368", + "accept": false, + "reason": "Both mention warnings, but they concern unrelated behaviors and models." + }, + { + "left": "issue:42886", + "right": "issue:43381", + "accept": false, + "reason": "Offline cache loading failure vs gradient checkpointing in eval mode are unrelated." + }, + { + "left": "issue:42907", + "right": "issue:45753", + "accept": false, + "reason": "Saving dequantized models and an infinite-loop inference bug are unrelated." + }, + { + "left": "issue:42907", + "right": "issue:44655", + "accept": false, + "reason": "Both mention saving, but one is dequantized model saving and the other is Pipeline.save_pretrained() support." + }, + { + "left": "issue:44380", + "right": "issue:45479", + "accept": false, + "reason": "Attention-scaling backend bug vs degenerate zero-loss classification issue; no shared code-path." + }, + { + "left": "issue:43352", + "right": "issue:43976", + "accept": false, + "reason": "Different regressions: Flash Attention support for a model vs Python version compatibility." + }, + { + "left": "issue:43064", + "right": "issue:44380", + "accept": false, + "reason": "Distributed training optimizer-state corruption is unrelated to GPT2 attention scaling under SDPA/FlashAttention." + }, + { + "left": "issue:43208", + "right": "issue:43526", + "accept": false, + "reason": "Different model bugs: xLSTM training issues vs BeitImageProcessorFast label reduction behavior." + }, + { + "left": "issue:44483", + "right": "issue:44514", + "accept": false, + "reason": "API request handling in chat/completions is unrelated to Qwen2_5_VLProcessor batching/padding crashes." + }, + { + "left": "issue:36683", + "right": "issue:43582", + "accept": false, + "reason": "Gemma3Config missing vocab_size and an Apple Silicon allocator TypeError are unrelated." + }, + { + "left": "issue:44222", + "right": "issue:45478", + "accept": false, + "reason": "Same model family, but different directions and failures: FP8 save_pretrained vs from_pretrained error on a newer version." + }, + { + "left": "issue:42490", + "right": "issue:43012", + "accept": false, + "reason": "Local path trailing-slash import warning is unrelated to the bfloat16 compile precision warning." + }, + { + "left": "issue:43618", + "right": "issue:43761", + "accept": false, + "reason": "Both are CLIP regressions, but one is missing attentions assignment and the other is hidden_states not returned; not the same concrete bug." + }, + { + "left": "issue:43012", + "right": "issue:44683", + "accept": false, + "reason": "Different compilation problems: a warning about precision vs a flex_attention failure on torch>=2.9." + }, + { + "left": "issue:42757", + "right": "issue:43824", + "accept": false, + "reason": "Importing is_offline_mode from huggingface_hub is unrelated to a missing Qwen2_5_VLForConditionalGeneration import." + }, + { + "left": "issue:43618", + "right": "issue:44380", + "accept": false, + "reason": "Different model/code paths: CLIP output fields vs GPT2 attention scaling under backend attention implementations." + }, + { + "left": "issue:38617", + "right": "issue:43502", + "accept": false, + "reason": "Distinct regressions: a missing configuration_utils symbol vs network calls despite local_files_only=True." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 43, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:36010", + "issue:36683", + "issue:41669", + "issue:42757", + "issue:42890", + "issue:42994", + "issue:43010", + "issue:43064", + "issue:43066", + "issue:43299", + "issue:43352", + "issue:43475", + "issue:43525", + "issue:43611", + "issue:43618", + "issue:43756", + "issue:43867", + "issue:43901", + "issue:43931", + "issue:43937", + "issue:43976", + "issue:43992", + "issue:44038", + "issue:44077", + "issue:44188", + "issue:44230", + "issue:44242", + "issue:44246", + "issue:44263", + "issue:44355", + "issue:44380", + "issue:44483", + "issue:44485", + "issue:44492", + "issue:44655", + "issue:44855", + "issue:44938", + "issue:45071", + "issue:45216", + "issue:45278", + "issue:45290", + "issue:45310", + "issue:45375", + "issue:45381", + "issue:45440", + "issue:45561", + "issue:45593", + "issue:45753" + ], + "soft_pairs": [ + "issue:44483|issue:45290", + "issue:44038|issue:44230", + "issue:43756|issue:44485", + "issue:44483|issue:45381", + "issue:43299|issue:43867", + "issue:43299|issue:45753", + "issue:43901|issue:44655", + "issue:36683|issue:43475", + "issue:42994|issue:44655", + "issue:43352|issue:43937", + "issue:44077|issue:44380", + "issue:43867|issue:45216", + "issue:36010|issue:42757", + "issue:43066|issue:45381", + "issue:36683|issue:44077", + "issue:43611|issue:44655", + "issue:43976|issue:44938", + "issue:36683|issue:43525", + "issue:43352|issue:45593", + "issue:44263|issue:45440", + "issue:43618|issue:45216", + "issue:36683|issue:44242", + "issue:44188|issue:44380", + "issue:43064|issue:44492", + "issue:43352|issue:43931", + "issue:43010|issue:45278", + "issue:43992|issue:45310", + "issue:43867|issue:45375", + "issue:41669|issue:44246", + "issue:44355|issue:44855", + "issue:36683|issue:43064", + "issue:42890|issue:45561", + "issue:44938|issue:45071" + ], + "prepared_review_unit_hash": "890124d5fae0451df18f893dd9409c3eaf8da290a044502f5a168d2b43132670", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 33, + "serialized_chars": 37171, + "estimated_input_tokens": 9293, + "estimated_eval_tokens": 18842 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is a mixed bag of Transformers v5 regressions and model-specific bugs, not a single duplicate set. The shared wording mostly reflects broad themes like loading/import/config breakage, but the concrete failures differ too much for PR-style duplicate merging.", + "confidence": 0.96, + "canonical_issue_reason": "issue:45278 is the broadest umbrella here: it explicitly captures widespread import breakage after upgrading Transformers, which fits the cluster\u2019s overall regression pattern better than the more narrow model-specific reports.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45278 is the best global representative because it is the most general and user-facing summary of the upgrade regression theme in this cluster.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44483", + "right": "issue:45290", + "accept": false, + "reason": "Different failures: chat/completions request acceptance vs apply_chat_template crashing on assistant tool-call messages." + }, + { + "left": "issue:44038", + "right": "issue:44230", + "accept": false, + "reason": "One is a Qwen3-VL-Moe loading bug; the other is fp8 inference support for Qwen3-VL/Qwen3.5 MoE, so the code paths and fixes differ." + }, + { + "left": "issue:43756", + "right": "issue:44485", + "accept": false, + "reason": "Both mention RoPE, but one is a Smollm3 layer-selection bug and the other is GLM-5 RoPE implementation discussion; not the same defect." + }, + { + "left": "issue:44483", + "right": "issue:45381", + "accept": false, + "reason": "Chat/completions request validation is unrelated to Qwen2.5-VL video vision_position_ids handling." + }, + { + "left": "issue:43299", + "right": "issue:43867", + "accept": false, + "reason": "Qwen3-VL-Moe loading regression and a generic sorted-state_dict loading error are different model-load problems." + }, + { + "left": "issue:43299", + "right": "issue:45753", + "accept": false, + "reason": "Model loading breakage and an infinite-loop-on-specific-image report are unrelated symptoms and likely different fixes." + }, + { + "left": "issue:43901", + "right": "issue:44655", + "accept": false, + "reason": "Docs mismatch for return_all_scores is not the same as the pipeline save_pretrained serialization bug." + }, + { + "left": "issue:36683", + "right": "issue:43475", + "accept": false, + "reason": "Gemma3 missing vocab_size and Sam3VisionOutput missing fpn_position_embeddings are different model attribute regressions." + }, + { + "left": "issue:42994", + "right": "issue:44655", + "accept": false, + "reason": "Quantized model saving failure and pipeline object save_pretrained failure affect different save paths." + }, + { + "left": "issue:43352", + "right": "issue:43937", + "accept": false, + "reason": "Nemotron FlashAttention2 support gating and GLM-5 GenerationConfig validation are unrelated." + }, + { + "left": "issue:44077", + "right": "issue:44380", + "accept": false, + "reason": "One is a patchtsmixer post_init policy change; the other is GPT2 attention scaling being ignored under SDPA/FlashAttention." + }, + { + "left": "issue:43867", + "right": "issue:45216", + "accept": false, + "reason": "Generic state_dict ordering load error and Qwen3.5 save_pretrained regression are different load/save bugs." + }, + { + "left": "issue:36010", + "right": "issue:42757", + "accept": false, + "reason": "Importing GenerationMixin from transformers.generation and importing is_offline_mode from huggingface_hub are distinct import regressions." + }, + { + "left": "issue:43066", + "right": "issue:45381", + "accept": false, + "reason": "Tokenizer decoder type in v5 and Qwen2.5-VL video position_ids are unrelated." + }, + { + "left": "issue:36683", + "right": "issue:44077", + "accept": false, + "reason": "Gemma3 config attribute missing and patchtsmixer post_init restriction have no shared code path." + }, + { + "left": "issue:43611", + "right": "issue:44655", + "accept": false, + "reason": "Base_model_prefix loading regression is not the same as pipeline serialization with save_pretrained." + }, + { + "left": "issue:43976", + "right": "issue:44938", + "accept": false, + "reason": "Python version compatibility regression and Python 3.14 load failure are too broad to treat as the same bug." + }, + { + "left": "issue:36683", + "right": "issue:43525", + "accept": false, + "reason": "Gemma3 missing vocab_size and Llama4 missing pad_token_id are separate config-attribute regressions." + }, + { + "left": "issue:43352", + "right": "issue:45593", + "accept": false, + "reason": "FlashAttention2 support for Nemotron and D-FINE auxiliary-loss behavior are unrelated issues." + }, + { + "left": "issue:44263", + "right": "issue:45440", + "accept": false, + "reason": "torch.split return handling in GlmMoeDsaIndexer is not the same as DeepseekV3MoE diverging from remote implementation." + }, + { + "left": "issue:43618", + "right": "issue:45216", + "accept": false, + "reason": "Missing CLIPOutput attentions and Qwen3.5 save_pretrained corruption are different regressions." + }, + { + "left": "issue:36683", + "right": "issue:44242", + "accept": false, + "reason": "Gemma3 vocab_size missing is unrelated to router-logit load-balancing loss behavior." + }, + { + "left": "issue:44188", + "right": "issue:44380", + "accept": false, + "reason": "Torch.compile attention-kernel divergence and GPT2 attention scaling ignored in SDPA/FlashAttention are different backend bugs." + }, + { + "left": "issue:43064", + "right": "issue:44492", + "accept": false, + "reason": "Distributed optimizer-state corruption and a cache-strategy typo are not the same problem." + }, + { + "left": "issue:43352", + "right": "issue:43931", + "accept": false, + "reason": "FlashAttention2 unsupported for Nemotron and Qwen3-VL weight-shape mismatch are unrelated model-loading/runtime failures." + }, + { + "left": "issue:43010", + "right": "issue:45278", + "accept": false, + "reason": "no_grad decoration on cache updates is unrelated to broad import errors after upgrading to v5.5.0." + }, + { + "left": "issue:43992", + "right": "issue:45310", + "accept": false, + "reason": "UMT5 missing embed_tokens.weight and Qwen3.5 MoE from_pretrained errors are different model-load regressions." + }, + { + "left": "issue:43867", + "right": "issue:45375", + "accept": false, + "reason": "Generic sorted-state_dict loading error and missing deepstack_visual_indexes field are distinct issues." + }, + { + "left": "issue:41669", + "right": "issue:44246", + "accept": false, + "reason": "44246 is a vague slow-import report, but 41669 is a specific import-* removal/performance fix; insufficient evidence they are the same concrete change." + }, + { + "left": "issue:42890", + "right": "issue:45561", + "accept": false, + "reason": "SAM HQ set_seed flakiness is unrelated to xdist races on captured_info.txt." + }, + { + "left": "issue:44938", + "right": "issue:45071", + "accept": false, + "reason": "Python 3.14 load failure and PretrainedConfig type-checking breakage are different release regressions, not one underlying bug." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 44, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:28282", + "issue:30333", + "issue:30990", + "issue:31795", + "issue:33453", + "issue:36296", + "issue:42490", + "issue:42548", + "issue:42831", + "issue:42890", + "issue:43010", + "issue:43065", + "issue:43232", + "issue:43299", + "issue:43502", + "issue:43575", + "issue:43618", + "issue:43723", + "issue:43867", + "issue:43874", + "issue:43976", + "issue:43992", + "issue:44062", + "issue:44075", + "issue:44077", + "issue:44112", + "issue:44222", + "issue:44351", + "issue:44355", + "issue:44360", + "issue:44387", + "issue:44484", + "issue:44485", + "issue:44556", + "issue:44749", + "issue:44857", + "issue:44938", + "issue:44964", + "issue:44977", + "issue:45125", + "issue:45341", + "issue:45357", + "issue:45375", + "issue:45381", + "issue:45440", + "issue:45478", + "issue:45600", + "issue:45850" + ], + "soft_pairs": [ + "issue:43867|issue:45357", + "issue:43010|issue:45850", + "issue:43992|issue:45478", + "issue:43010|issue:43976", + "issue:44938|issue:45381", + "issue:43618|issue:44077", + "issue:42490|issue:44075", + "issue:36296|issue:43575", + "issue:44360|issue:45440", + "issue:28282|issue:33453", + "issue:43867|issue:44484", + "issue:43618|issue:44857", + "issue:43618|issue:44977", + "issue:42890|issue:44964", + "issue:42890|issue:44112", + "issue:42890|issue:45600", + "issue:43502|issue:44062", + "issue:43723|issue:44556", + "issue:42890|issue:45341", + "issue:44222|issue:45125", + "issue:44387|issue:44556", + "issue:42548|issue:44351", + "issue:43010|issue:43065", + "issue:43502|issue:44485", + "issue:44355|issue:44749", + "issue:45125|issue:45478", + "issue:43010|issue:43232", + "issue:30990|issue:31795", + "issue:45125|issue:45375", + "issue:30333|issue:33453", + "issue:43299|issue:44556", + "issue:43502|issue:44222", + "issue:42831|issue:43874" + ], + "prepared_review_unit_hash": "a421655fb8e437156fc30156ba9116792704a22b9552940f22730118d06d0655", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 33, + "serialized_chars": 37013, + "estimated_input_tokens": 9254, + "estimated_eval_tokens": 18764 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a loose set of Transformers issues sharing broad loading/version-regression language, but the concrete bugs are different enough that none of the soft pairs should be merged as duplicates.", + "confidence": 0.53, + "canonical_issue_reason": "Issue 45850 is the broadest open umbrella here: it describes widespread breakage across many models after 5.0, making it the best representative for a mixed regression cluster.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 45850 is the best global issue candidate because it is open, generic, and can plausibly collect related regression reports without being tied to one model-specific failure.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43867", + "right": "issue:45357", + "accept": false, + "reason": "Both involve model loading/saving regressions, but one is about sorted state_dict loading and the other about incorrect visual encoder keys in save_pretrained." + }, + { + "left": "issue:43010", + "right": "issue:45850", + "accept": false, + "reason": "Decorating cache/layer update with no_grad is a specific API fix, not the broad multi-model regression described in 45850." + }, + { + "left": "issue:43992", + "right": "issue:45478", + "accept": false, + "reason": "Both mention from_pretrained failures, but they target different Qwen/UMT5 model paths and likely different missing weights." + }, + { + "left": "issue:43010", + "right": "issue:43976", + "accept": false, + "reason": "One is a no_grad behavior bug; the other is a Python-version compatibility failure." + }, + { + "left": "issue:44938", + "right": "issue:45381", + "accept": false, + "reason": "Python 3.14 import/load failure and Qwen2.5-VL vision_position_ids corruption are unrelated code paths." + }, + { + "left": "issue:43618", + "right": "issue:44077", + "accept": false, + "reason": "CLIPOutput attentions assignment and PatchTSMixer post_init validation are unrelated model API issues." + }, + { + "left": "issue:42490", + "right": "issue:44075", + "accept": false, + "reason": "A local-path import warning and SGD optimizer args being ignored are unrelated." + }, + { + "left": "issue:36296", + "right": "issue:43575", + "accept": false, + "reason": "Both mention tensor parallelism, but one is a training bug and the other is an OOM when loading a specific Qwen2 checkpoint." + }, + { + "left": "issue:44360", + "right": "issue:45440", + "accept": false, + "reason": "DSA indexer ReLU discussion and DeepSeekV3 implementation divergence are different components and problems." + }, + { + "left": "issue:28282", + "right": "issue:33453", + "accept": false, + "reason": "PyTorch missing ImportError and tokenizer loading regression are separate failure modes." + }, + { + "left": "issue:43867", + "right": "issue:44484", + "accept": false, + "reason": "Both touch save/load APIs, but one is a loading bug and the other is a max_shard_size behavior question." + }, + { + "left": "issue:43618", + "right": "issue:44857", + "accept": false, + "reason": "CLIPOutput attentions and float16 AMP crash in LwDetrImageLoss are unrelated." + }, + { + "left": "issue:43618", + "right": "issue:44977", + "accept": false, + "reason": "A missing attentions field is not the same as Qwen3.5 flash-attention generation corruption." + }, + { + "left": "issue:42890", + "right": "issue:44964", + "accept": false, + "reason": "A flaky test due to missing set_seed is unrelated to a Phi-4 multimodal model loading regression." + }, + { + "left": "issue:42890", + "right": "issue:44112", + "accept": false, + "reason": "Both are test-related, but they fail for different reasons and on different model suites." + }, + { + "left": "issue:42890", + "right": "issue:45600", + "accept": false, + "reason": "A seed-sensitive integration test is unrelated to auto_mappings referencing removed Sam3LiteText configs." + }, + { + "left": "issue:43502", + "right": "issue:44062", + "accept": false, + "reason": "Local_files_only network leakage and AddedToken keyword collision are different tokenizer/runtime bugs." + }, + { + "left": "issue:43723", + "right": "issue:44556", + "accept": false, + "reason": "Both are loading regressions, but one is tokenizer-specific and the other is checkpoint reload after version upgrade." + }, + { + "left": "issue:42890", + "right": "issue:45341", + "accept": false, + "reason": "A flaky model test and a testing_utils bug are not the same underlying issue." + }, + { + "left": "issue:44222", + "right": "issue:45125", + "accept": false, + "reason": "FP8 save_pretrained and missing _tp_plan for tensor parallelism affect different code paths." + }, + { + "left": "issue:44387", + "right": "issue:44556", + "accept": false, + "reason": "CUDA memory growth under int4 quantization is unrelated to checkpoint reload breakage after upgrade." + }, + { + "left": "issue:42548", + "right": "issue:44351", + "accept": false, + "reason": "Both are import errors, but they concern different missing symbols and likely different export regressions." + }, + { + "left": "issue:43010", + "right": "issue:43065", + "accept": false, + "reason": "A no_grad decoration bug and a dummy Conv2d in Sam3PixelDecoder are unrelated." + }, + { + "left": "issue:43502", + "right": "issue:44485", + "accept": false, + "reason": "Network access despite local_files_only and GLM-5 RoPE implementation are unrelated." + }, + { + "left": "issue:44355", + "right": "issue:44749", + "accept": false, + "reason": "Compiled Python file errors and slow filtering after upgrade are unrelated." + }, + { + "left": "issue:45125", + "right": "issue:45478", + "accept": false, + "reason": "They are both Qwen3.5 MoE loading issues, but the first is specifically missing _tp_plan while the second is a broader from_pretrained error without enough evidence of the same root cause." + }, + { + "left": "issue:43010", + "right": "issue:43232", + "accept": false, + "reason": "Cache/layer no_grad decoration and generation kwargs sync behavior are different generation internals." + }, + { + "left": "issue:30990", + "right": "issue:31795", + "accept": false, + "reason": "Sentence Transformers loading hang and forward-doc confusion are unrelated." + }, + { + "left": "issue:45125", + "right": "issue:45375", + "accept": false, + "reason": "Missing _tp_plan for Qwen3.5 MoE and missing deepstack_visual_indexes in Qwen3_5MoeVisionConfig are different configuration problems." + }, + { + "left": "issue:30333", + "right": "issue:33453", + "accept": false, + "reason": "MLflow reporting on failed training and tokenizer loading regression are unrelated." + }, + { + "left": "issue:43299", + "right": "issue:44556", + "accept": false, + "reason": "Both are version-related loading failures, but they affect different model families and likely different loading code paths." + }, + { + "left": "issue:43502", + "right": "issue:44222", + "accept": false, + "reason": "A local-files-only network bug and FP8 save_pretrained issues are unrelated." + }, + { + "left": "issue:42831", + "right": "issue:43874", + "accept": false, + "reason": "FineGrainedFP8 accuracy and GLM46V image-processor method absence are different model-specific bugs." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 45, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:30333", + "issue:31795", + "issue:33453", + "issue:34689", + "issue:36296", + "issue:38617", + "issue:41628", + "issue:41669", + "issue:42371", + "issue:42548", + "issue:42757", + "issue:43502", + "issue:43525", + "issue:43526", + "issue:43572", + "issue:43576", + "issue:43673", + "issue:43824", + "issue:43827", + "issue:43901", + "issue:44038", + "issue:44222", + "issue:44230", + "issue:44261", + "issue:44273", + "issue:44315", + "issue:44355", + "issue:44360", + "issue:44393", + "issue:44485", + "issue:44492", + "issue:44556", + "issue:44623", + "issue:44655", + "issue:44704", + "issue:44908", + "issue:44938", + "issue:44961", + "issue:45003", + "issue:45071", + "issue:45092", + "issue:45125", + "issue:45230", + "issue:45362", + "issue:45468", + "issue:45542", + "issue:45753", + "issue:45850" + ], + "soft_pairs": [ + "issue:44315|issue:44360", + "issue:43502|issue:44961", + "issue:44355|issue:45071", + "issue:30333|issue:36296", + "issue:44355|issue:44655", + "issue:43827|issue:44961", + "issue:44222|issue:44230", + "issue:33453|issue:34689", + "issue:42757|issue:43576", + "issue:33453|issue:36296", + "issue:43576|issue:43824", + "issue:43576|issue:45071", + "issue:31795|issue:38617", + "issue:44961|issue:45468", + "issue:41628|issue:42548", + "issue:38617|issue:43576", + "issue:43576|issue:44938", + "issue:43673|issue:44492", + "issue:44908|issue:44961", + "issue:44485|issue:44961", + "issue:36296|issue:43901", + "issue:41628|issue:44355", + "issue:36296|issue:43526", + "issue:45362|issue:45850", + "issue:44261|issue:44360", + "issue:44623|issue:45230", + "issue:44273|issue:45542", + "issue:44273|issue:45003", + "issue:44393|issue:45753", + "issue:44393|issue:45071", + "issue:44038|issue:45125", + "issue:43576|issue:44556", + "issue:43525|issue:45850", + "issue:41669|issue:43673", + "issue:44556|issue:45092", + "issue:44492|issue:45850", + "issue:43572|issue:45850", + "issue:42371|issue:44492", + "issue:44704|issue:44908" + ], + "prepared_review_unit_hash": "f45d1b0f4a8716cd6529002bf46596050cd19ee182ccedd680a8836fc92f822d", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 39, + "serialized_chars": 38478, + "estimated_input_tokens": 9620, + "estimated_eval_tokens": 19496 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is mostly a collection of unrelated Transformers v5-era regressions and import/loading complaints. Issue 45850 is the broadest umbrella, but most pairs are not true duplicates; they only share a generic theme like 'v5 broke something.'", + "confidence": 0.81, + "canonical_issue_reason": "issue:45850 is the broadest umbrella in the set and best matches the overall 'v5 breaks many models' regression theme.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45850 is the best single anchor for the cluster, though it is too vague to serve as a precise duplicate target.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44315", + "right": "issue:44360", + "accept": false, + "reason": "Different subsystems and symptoms: model_init/Liger Kernel vs DSA indexer/ReLU." + }, + { + "left": "issue:43502", + "right": "issue:44961", + "accept": false, + "reason": "Unrelated issues: offline/local_files_only behavior vs an empty 'racoon' report." + }, + { + "left": "issue:44355", + "right": "issue:45071", + "accept": false, + "reason": "One is about compiled Python file loading; the other is PretrainedConfig type checking." + }, + { + "left": "issue:30333", + "right": "issue:36296", + "accept": false, + "reason": "MLFlow reporting on failed training jobs is unrelated to tensor-parallel training behavior." + }, + { + "left": "issue:44355", + "right": "issue:44655", + "accept": false, + "reason": "Compiled Python file errors and pipeline save_pretrained failures are different bugs." + }, + { + "left": "issue:43827", + "right": "issue:44961", + "accept": false, + "reason": "Docs mentioning pipeline() after v5 removals is unrelated to the 'racoon' issue." + }, + { + "left": "issue:44222", + "right": "issue:44230", + "accept": false, + "reason": "Both mention FP8/MoE, but one is save_pretrained behavior and the other is inference support." + }, + { + "left": "issue:33453", + "right": "issue:34689", + "accept": false, + "reason": "Both are loading regressions, but they point to different code paths and a specific Llama Vision case." + }, + { + "left": "issue:42757", + "right": "issue:43576", + "accept": false, + "reason": "ImportError from huggingface_hub is unrelated to the v5 env command breakage." + }, + { + "left": "issue:33453", + "right": "issue:36296", + "accept": false, + "reason": "Tokenizer loading regression is unrelated to tensor-parallel training." + }, + { + "left": "issue:43576", + "right": "issue:43824", + "accept": false, + "reason": "Broken env command and missing Qwen2.5-VL export are separate failures." + }, + { + "left": "issue:43576", + "right": "issue:45071", + "accept": false, + "reason": "Env command breakage is unrelated to PretrainedConfig type-checking regressions." + }, + { + "left": "issue:31795", + "right": "issue:38617", + "accept": false, + "reason": "Documentation confusion in forward args is unrelated to an ImportError for layer_type_validation." + }, + { + "left": "issue:44961", + "right": "issue:45468", + "accept": false, + "reason": "A placeholder issue and Gemma audio positional encoding bug are not duplicates." + }, + { + "left": "issue:41628", + "right": "issue:42548", + "accept": false, + "reason": "Both are import errors, but for different symbols and likely different export gaps." + }, + { + "left": "issue:38617", + "right": "issue:43576", + "accept": false, + "reason": "ImportError for layer_type_validation and env command breakage are different problems." + }, + { + "left": "issue:43576", + "right": "issue:44938", + "accept": false, + "reason": "Broken env command is not the same as Python 3.14 load failures." + }, + { + "left": "issue:43673", + "right": "issue:44492", + "accept": false, + "reason": "Chunked_prefill cache issue and a cache-strategy typo are not the same bug." + }, + { + "left": "issue:44908", + "right": "issue:44961", + "accept": false, + "reason": "Scheduler kwargs handling is unrelated to the placeholder 'racoon' issue." + }, + { + "left": "issue:44485", + "right": "issue:44961", + "accept": false, + "reason": "RoPE implementation discussion is unrelated to the placeholder issue." + }, + { + "left": "issue:36296", + "right": "issue:43901", + "accept": false, + "reason": "Tensor-parallel training bug is unrelated to documentation about return_all_scores." + }, + { + "left": "issue:41628", + "right": "issue:44355", + "accept": false, + "reason": "AutoImageProcessor import error and compiled-file loading errors are different regressions." + }, + { + "left": "issue:36296", + "right": "issue:43526", + "accept": false, + "reason": "Tensor parallel training and BeitImageProcessorFast label reduction are unrelated." + }, + { + "left": "issue:45362", + "right": "issue:45850", + "accept": false, + "reason": "Qwen3.5 chat crash is a specific model/runtime bug; 45850 is only a broad umbrella." + }, + { + "left": "issue:44261", + "right": "issue:44360", + "accept": false, + "reason": "MLA rms_norm_eps precision issue is unrelated to the DSA/ReLU report." + }, + { + "left": "issue:44623", + "right": "issue:45230", + "accept": false, + "reason": "Missing processor.save_pretrained files is unrelated to the placeholder bug report." + }, + { + "left": "issue:44273", + "right": "issue:45542", + "accept": false, + "reason": "Lazy loading problems and tensorboard/TensorFlow backend errors are different failures." + }, + { + "left": "issue:44273", + "right": "issue:45003", + "accept": false, + "reason": "Lazy loading behavior and unsafe sys.modules access are related only broadly, not as the same concrete bug." + }, + { + "left": "issue:44393", + "right": "issue:45753", + "accept": false, + "reason": "Different Qwen model/version issues: 2D box output vs infinite loop on one image." + }, + { + "left": "issue:44393", + "right": "issue:45071", + "accept": false, + "reason": "Qwen3-VL output bugs are unrelated to PretrainedConfig type-checking." + }, + { + "left": "issue:44038", + "right": "issue:45125", + "accept": false, + "reason": "Generic Qwen3-VL-MoE bug report and missing _tp_plan for tensor parallelism are distinct." + }, + { + "left": "issue:43576", + "right": "issue:44556", + "accept": false, + "reason": "Env command breakage is unrelated to checkpoint reload incompatibility after upgrade." + }, + { + "left": "issue:43525", + "right": "issue:45850", + "accept": false, + "reason": "Missing pad_token_id in Llama4Config is too specific to be the same as the broad v5 breakage umbrella." + }, + { + "left": "issue:44492", + "right": "issue:45850", + "accept": false, + "reason": "Cache strategy typo is not the same as the broad 'v5 breaks too many models' complaint." + }, + { + "left": "issue:43572", + "right": "issue:45850", + "accept": false, + "reason": "StableLmConfig pad_token_idx regression is a separate model-specific issue." + }, + { + "left": "issue:42371", + "right": "issue:44492", + "accept": false, + "reason": "TF32 behavior settings and cache strategy typo are unrelated." + }, + { + "left": "issue:44704", + "right": "issue:44908", + "accept": false, + "reason": "Missing kwargs in AutoProcessor.from_pretrained is a different code path from scheduler kwargs handling." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 46, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:31795", + "issue:33453", + "issue:34689", + "issue:37428", + "issue:41669", + "issue:42371", + "issue:42890", + "issue:42913", + "issue:43066", + "issue:43352", + "issue:43408", + "issue:43502", + "issue:43519", + "issue:43576", + "issue:43606", + "issue:43704", + "issue:43827", + "issue:43874", + "issue:43976", + "issue:44077", + "issue:44246", + "issue:44261", + "issue:44263", + "issue:44273", + "issue:44295", + "issue:44297", + "issue:44315", + "issue:44351", + "issue:44360", + "issue:44393", + "issue:44485", + "issue:44492", + "issue:44623", + "issue:44655", + "issue:44829", + "issue:44861", + "issue:44908", + "issue:44945", + "issue:44998", + "issue:45071", + "issue:45092", + "issue:45230", + "issue:45278", + "issue:45310", + "issue:45341", + "issue:45468", + "issue:45478", + "issue:45850" + ], + "soft_pairs": [ + "issue:42890|issue:44908", + "issue:45278|issue:45850", + "issue:44246|issue:44273", + "issue:43408|issue:44315", + "issue:43976|issue:44393", + "issue:43352|issue:44393", + "issue:43576|issue:44908", + "issue:44297|issue:45230", + "issue:44295|issue:45230", + "issue:44908|issue:45071", + "issue:41669|issue:44351", + "issue:41669|issue:45341", + "issue:44492|issue:44945", + "issue:43874|issue:45850", + "issue:44393|issue:45478", + "issue:44077|issue:44861", + "issue:43827|issue:44908", + "issue:31795|issue:33453", + "issue:44315|issue:44829", + "issue:41669|issue:44908", + "issue:43576|issue:44273", + "issue:44655|issue:44861", + "issue:44360|issue:45468", + "issue:45230|issue:45310", + "issue:44273|issue:44908", + "issue:43519|issue:44485", + "issue:43519|issue:44908", + "issue:41669|issue:43704", + "issue:31795|issue:34689", + "issue:31795|issue:37428", + "issue:44315|issue:45092", + "issue:44485|issue:45468", + "issue:43704|issue:44485", + "issue:42371|issue:43704", + "issue:43502|issue:43519", + "issue:43519|issue:44623", + "issue:43519|issue:43704", + "issue:44263|issue:45468", + "issue:43606|issue:45468", + "issue:44261|issue:45468", + "issue:42913|issue:43066", + "issue:44908|issue:44998" + ], + "prepared_review_unit_hash": "d2b11dbba0b97c79375e58757a758ab2c19c1709d569f65a3a1961b5a71158dd", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 42, + "serialized_chars": 39635, + "estimated_input_tokens": 9909, + "estimated_eval_tokens": 20074 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a mix of Transformers v5 regression reports across tokenizers, imports, loading, and model-specific failures. issue:45278 remains the best umbrella issue, but the soft edges should stay conservative: none of these pairs are clearly the same concrete bug from the provided titles alone.", + "confidence": 0.58, + "canonical_issue_reason": "issue:45278 is still the strongest canonical issue: it is open and explicitly captures the broad multi-model breakage after the 4.57.0 -> 5.5.0 upgrade, making it a practical umbrella for this cluster.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45278 is the best global issue target because it is open, broad enough to subsume related v5 breakage reports, and specific enough to remain useful as a tracker.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:42890", + "right": "issue:44908", + "accept": false, + "reason": "SamHQ test flakiness from missing set_seed is unrelated to inverse_sqrt scheduler kwargs handling." + }, + { + "left": "issue:45278", + "right": "issue:45850", + "accept": false, + "reason": "Both are broad v5 breakage complaints, but they do not describe the same concrete failure mode." + }, + { + "left": "issue:44246", + "right": "issue:44273", + "accept": false, + "reason": "Import slowness and lazy-loading failure are related themes, but not clearly the same bug." + }, + { + "left": "issue:43408", + "right": "issue:44315", + "accept": false, + "reason": "Model-type warning for sam3_tracker is unrelated to Liger Kernel application in model_init." + }, + { + "left": "issue:43976", + "right": "issue:44393", + "accept": false, + "reason": "Python version compatibility and Qwen3-VL bbox output errors are unrelated." + }, + { + "left": "issue:43352", + "right": "issue:44393", + "accept": false, + "reason": "Flash Attention 2 support error is unrelated to Qwen3-VL bounding-box hallucination/output bugs." + }, + { + "left": "issue:43576", + "right": "issue:44908", + "accept": false, + "reason": "Broken transformers env command is unrelated to scheduler kwargs being ignored." + }, + { + "left": "issue:44297", + "right": "issue:45230", + "accept": false, + "reason": "Specific tokenizer_class mismatch is too different from a generic bug report to treat as the same issue." + }, + { + "left": "issue:44295", + "right": "issue:45230", + "accept": false, + "reason": "Position_ids buffer access error is not the same as an unspecified generic bug report." + }, + { + "left": "issue:44908", + "right": "issue:45071", + "accept": false, + "reason": "Scheduler kwargs bug and PretrainedConfig type-checking regression are different code paths." + }, + { + "left": "issue:41669", + "right": "issue:44351", + "accept": false, + "reason": "Import-* performance regression is unrelated to the HybridCache import error." + }, + { + "left": "issue:41669", + "right": "issue:45341", + "accept": false, + "reason": "Import-* performance regression is unrelated to a testing_utils bug." + }, + { + "left": "issue:44492", + "right": "issue:44945", + "accept": false, + "reason": "Cache-strategy typo and pipeline-parallelism output corruption are unrelated." + }, + { + "left": "issue:43874", + "right": "issue:45850", + "accept": false, + "reason": "Missing get_number_of_image_patches is a concrete GLM46V bug, not the same as generic model breakage." + }, + { + "left": "issue:44393", + "right": "issue:45478", + "accept": false, + "reason": "Qwen3-VL bbox issues and Qwen3.5 MoE from_pretrained errors are different failures." + }, + { + "left": "issue:44077", + "right": "issue:44861", + "accept": false, + "reason": "patchtsmixer post_init policy and tied-weight-key AttributeError are unrelated." + }, + { + "left": "issue:43827", + "right": "issue:44908", + "accept": false, + "reason": "Docs still referencing pipeline() is unrelated to inverse_sqrt scheduler kwargs." + }, + { + "left": "issue:31795", + "right": "issue:33453", + "accept": false, + "reason": "Confusing forward docs and tokenizer loading regression are different issues." + }, + { + "left": "issue:44315", + "right": "issue:44829", + "accept": false, + "reason": "Liger Kernel not applied with model_init is unrelated to flash_attention_3 degenerate training." + }, + { + "left": "issue:41669", + "right": "issue:44908", + "accept": false, + "reason": "Import-* performance regression is unrelated to a scheduler keyword bug." + }, + { + "left": "issue:43576", + "right": "issue:44273", + "accept": false, + "reason": "Broken env CLI and lazy-loading failure are only loosely related." + }, + { + "left": "issue:44655", + "right": "issue:44861", + "accept": false, + "reason": "Pipeline save_pretrained failure is unrelated to _tied_weight_keys list handling." + }, + { + "left": "issue:44360", + "right": "issue:45468", + "accept": false, + "reason": "DSA indexer ReLU issue and Gemma-4 audio positional encoding bug are different model-specific problems." + }, + { + "left": "issue:45230", + "right": "issue:45310", + "accept": false, + "reason": "A generic bug report is too vague to confirm it is the same as the Qwen3.5 MoE from_pretrained error." + }, + { + "left": "issue:44273", + "right": "issue:44908", + "accept": false, + "reason": "Lazy loading and inverse_sqrt scheduler kwargs are unrelated." + }, + { + "left": "issue:43519", + "right": "issue:44485", + "accept": false, + "reason": "Qwen3VL timestamp calculation and GLM-5 RoPE implementation are different model bugs." + }, + { + "left": "issue:43519", + "right": "issue:44908", + "accept": false, + "reason": "Timestamp calculation in Qwen3VL is unrelated to scheduler kwargs handling." + }, + { + "left": "issue:41669", + "right": "issue:43704", + "accept": false, + "reason": "Import-* performance regression is unrelated to a VRAM leak in dataloader threads." + }, + { + "left": "issue:31795", + "right": "issue:34689", + "accept": false, + "reason": "Model.forward documentation confusion is unrelated to a Llama 3.2 model-loading regression." + }, + { + "left": "issue:31795", + "right": "issue:37428", + "accept": false, + "reason": "Forward-doc confusion is unrelated to the flash-attention import error." + }, + { + "left": "issue:44315", + "right": "issue:45092", + "accept": false, + "reason": "Both involve model initialization, but they describe different concrete incompatibilities." + }, + { + "left": "issue:44485", + "right": "issue:45468", + "accept": false, + "reason": "Different models and different bugs: GLM-5 RoPE versus Gemma-4 audio positional encoding." + }, + { + "left": "issue:43704", + "right": "issue:44485", + "accept": false, + "reason": "VRAM leak in multi-threaded dataloading is unrelated to GLM-5 RoPE issues." + }, + { + "left": "issue:42371", + "right": "issue:43704", + "accept": false, + "reason": "TF32 configuration guidance is unrelated to a VRAM leak." + }, + { + "left": "issue:43502", + "right": "issue:43519", + "accept": false, + "reason": "local_files_only network requests and Qwen3VL timestamp calculation are unrelated." + }, + { + "left": "issue:43519", + "right": "issue:44623", + "accept": false, + "reason": "Timestamp calculation bug and processor.save_pretrained missing files are unrelated." + }, + { + "left": "issue:43519", + "right": "issue:43704", + "accept": false, + "reason": "Timestamp calculation bug and VRAM leak are unrelated." + }, + { + "left": "issue:44263", + "right": "issue:45468", + "accept": false, + "reason": "torch.split return-values issue and Gemma-4 audio positional encoding bug are unrelated." + }, + { + "left": "issue:43606", + "right": "issue:45468", + "accept": false, + "reason": "CPU offload device mismatch and Gemma-4 audio encoding issue are unrelated." + }, + { + "left": "issue:44261", + "right": "issue:45468", + "accept": false, + "reason": "LayerNorm epsilon precision issue and Gemma-4 audio encoding issue are unrelated." + }, + { + "left": "issue:42913", + "right": "issue:43066", + "accept": false, + "reason": "The titles point to two different tokenizer-v5 regressions: one is a broad behavior difference, the other is a specific wrong decoder-type bug. That is not enough to confirm the same underlying issue." + }, + { + "left": "issue:44908", + "right": "issue:44998", + "accept": false, + "reason": "Scheduler kwargs bug is unrelated to the placeholder/unemployment issue." + } + ] + }, + "evaluator_result": { + "accept": false, + "feedback": "The cluster summary is mostly grounded, and the 45278 umbrella choice is plausible, but I can\u2019t accept the output because one soft-edge verdict is too strong: 42913 \u2194 43066 is marked true even though the packet only gives titles showing two different tokenizer-v5 regressions (general behavior difference vs wrong decoder type), which is not enough to confirm the same bug. That makes the overall judgment insufficiently conservative." + }, + "evaluator_used": true, + "retried": true, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 1, + "component_count": 17, + "review_unit_index": 47, + "review_unit_count": 47, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-10000-400", + "nodes": [ + "issue:30333", + "issue:31515", + "issue:31795", + "issue:33453", + "issue:36296", + "issue:38617", + "issue:43298", + "issue:43525", + "issue:43572", + "issue:43618", + "issue:44230", + "issue:44492", + "issue:44623", + "issue:44861", + "issue:44998", + "issue:45290", + "issue:45310", + "issue:45341", + "issue:45850" + ], + "soft_pairs": [ + "issue:44861|issue:45290", + "issue:44492|issue:44998", + "issue:44998|issue:45850", + "issue:43618|issue:44998", + "issue:44998|issue:45341", + "issue:43525|issue:44861", + "issue:43572|issue:44861", + "issue:31515|issue:38617", + "issue:44230|issue:44623", + "issue:30333|issue:31515", + "issue:31515|issue:36296", + "issue:31515|issue:31795", + "issue:44230|issue:45310", + "issue:31515|issue:33453", + "issue:43298|issue:44230" + ], + "prepared_review_unit_hash": "111f8039cce9a3727b6939c68c609b23c223a5569a22976c763bf66425840a97", + "component_budget": { + "node_count": 400, + "item_count": 400, + "soft_pair_count": 1499, + "serialized_chars": 772970, + "estimated_input_tokens": 193243, + "estimated_eval_tokens": 386742 + }, + "budget": { + "node_count": 19, + "item_count": 19, + "soft_pair_count": 15, + "serialized_chars": 15089, + "estimated_input_tokens": 3773, + "estimated_eval_tokens": 7802 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is heterogeneous: it mixes unrelated regressions, a docs typo, a model-config attribute bug, and several loading/runtime issues. None of the soft pairs look like the same underlying bug, so I would not merge any of them as duplicates.", + "confidence": 0.19, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": "Issue 45310 is the most concrete and actionable regression report in the set (a specific from_pretrained failure on Qwen3.5 MoE), but the surrounding issues are too diverse for a true canonical duplicate.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44861", + "right": "issue:45290", + "accept": false, + "reason": "Different failures: tied-weight key handling vs chat template parsing with tool calls. No shared code-path problem." + }, + { + "left": "issue:44492", + "right": "issue:44998", + "accept": false, + "reason": "A cache-strategy typo is unrelated to an unemployment report." + }, + { + "left": "issue:44998", + "right": "issue:45850", + "accept": false, + "reason": "Generic broad complaint about many models breaking is not the same as an unemployment issue." + }, + { + "left": "issue:43618", + "right": "issue:44998", + "accept": false, + "reason": "CLIP attentions assignment regression is unrelated to the unemployment report." + }, + { + "left": "issue:44998", + "right": "issue:45341", + "accept": false, + "reason": "Unrelated subjects: unemployment vs a testing_utils bug." + }, + { + "left": "issue:43525", + "right": "issue:44861", + "accept": false, + "reason": "Llama4Config pad_token_id absence and tied-weight key crash are different model/config issues." + }, + { + "left": "issue:43572", + "right": "issue:44861", + "accept": false, + "reason": "StableLmConfig pad_token_idx regression is not the same bug as _get_tied_weight_keys crashing." + }, + { + "left": "issue:31515", + "right": "issue:38617", + "accept": false, + "reason": "Slow from_pretrained checkpoint loading and an ImportError for layer_type_validation are unrelated." + }, + { + "left": "issue:44230", + "right": "issue:44623", + "accept": false, + "reason": "FP8 Qwen/VL support request is different from processor.save_pretrained missing files." + }, + { + "left": "issue:30333", + "right": "issue:31515", + "accept": false, + "reason": "MLFlow job-status reporting and checkpoint loading performance are unrelated." + }, + { + "left": "issue:31515", + "right": "issue:36296", + "accept": false, + "reason": "Checkpoint loading slowness is not the same as a tensor-parallel training bug." + }, + { + "left": "issue:31515", + "right": "issue:31795", + "accept": false, + "reason": "Loading speed regression and documentation confusion are different classes of issue." + }, + { + "left": "issue:44230", + "right": "issue:45310", + "accept": false, + "reason": "Both mention Qwen/from_pretrained-like contexts, but one is FP8 support and the other is a specific loading error; not the same concrete bug." + }, + { + "left": "issue:31515", + "right": "issue:33453", + "accept": false, + "reason": "Tokenizer loading regression is distinct from slow checkpoint loading." + }, + { + "left": "issue:43298", + "right": "issue:44230", + "accept": false, + "reason": "Continuous batching audio-model support and FP8 Qwen/VL support are different feature gaps." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-13244-3", + "nodes": [ + "issue:13244", + "issue:15980", + "issue:16618" + ], + "soft_pairs": [ + "issue:13244|issue:16618", + "issue:15980|issue:16618" + ], + "prepared_review_unit_hash": "f4e96e0d13b14c57a4991de17e4204714c51358c4c93d0cefc666cfd89a07c13", + "component_budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 2, + "serialized_chars": 2478, + "estimated_input_tokens": 620, + "estimated_eval_tokens": 1496 + }, + "budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 2, + "serialized_chars": 2478, + "estimated_input_tokens": 620, + "estimated_eval_tokens": 1496 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The cluster only has loose topical similarity around model/token handling. Issue 13244 is about tokenization mismatch, 15980 is about an auth/download error message, and 16618 is a model loading failure. They do not look like duplicates of the same underlying bug.", + "confidence": 0.86, + "canonical_issue_reason": "issue:16618 is the most concrete user-facing runtime failure in the cluster and the broadest model-loading problem; the others are narrower and about different code paths.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:16618 is the best anchor for the cluster because it is the clearest concrete bug report, but it still does not subsume the other two issues.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:13244", + "right": "issue:16618", + "accept": false, + "reason": "Tokenization output differing from TensorFlow code is a different bug class from failing to load a model; the symptoms and likely code paths do not match." + }, + { + "left": "issue:15980", + "right": "issue:16618", + "accept": false, + "reason": "Bad error messaging for unauthenticated private-model download is not the same as a generic 'can't load bert-base-uncased' failure; related area, but not the same underlying defect." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 3, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-26350-2", + "nodes": [ + "issue:26350", + "issue:33260" + ], + "soft_pairs": [ + "issue:26350|issue:33260" + ], + "prepared_review_unit_hash": "3b7ac85596561b41adac32f95e86a3080e054d8dcc787e3bde3d11856dfe6f62", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1589, + "estimated_input_tokens": 398, + "estimated_eval_tokens": 1052 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1589, + "estimated_input_tokens": 398, + "estimated_eval_tokens": 1052 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are two separate community-contribution feature requests: one for Flash Attention 2 architecture support and one for GGUF architecture support. They share a broad theme but not the same underlying bug/change.", + "confidence": 0.98, + "canonical_issue_reason": "No canonical issue: the requests target different features and different code paths.", + "canonical_pr_reason": null, + "best_issue_reason": "Neither issue is a duplicate of the other, so there is no single canonical issue for the pair.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:26350", + "right": "issue:33260", + "accept": false, + "reason": "Different requested capabilities (Flash Attention 2 vs GGUF support); same broad subsystem, but not the same underlying change or bug." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 4, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-27088-4", + "nodes": [ + "issue:27088", + "issue:43834", + "issue:44365", + "issue:44370" + ], + "soft_pairs": [ + "issue:27088|issue:44370", + "issue:27088|issue:44365", + "issue:27088|issue:43834", + "issue:43834|issue:44370", + "issue:43834|issue:44365", + "issue:44365|issue:44370" + ], + "prepared_review_unit_hash": "1dff98e9c7c67da2e0c92448858e5db7da25cc236f3d9605926697bc785fafb8", + "component_budget": { + "node_count": 4, + "item_count": 4, + "soft_pair_count": 6, + "serialized_chars": 4452, + "estimated_input_tokens": 1113, + "estimated_eval_tokens": 2482 + }, + "budget": { + "node_count": 4, + "item_count": 4, + "soft_pair_count": 6, + "serialized_chars": 4452, + "estimated_input_tokens": 1113, + "estimated_eval_tokens": 2482 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These look like translation-request issues, but only one is concrete (Turkish). The others are generic template-style placeholders and may represent different language requests, so I would not collapse them as duplicates.", + "confidence": 0.74, + "canonical_issue_reason": "Issue 27088 is the most concrete and specific: it names Turkish explicitly and looks like a real translation request rather than a template placeholder.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 27088 is the strongest representative because it is specific, actionable, and not just a generic language-template title.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:27088", + "right": "issue:44370", + "accept": false, + "reason": "Turkish-specific request vs a generic language-template issue; the similarity appears to come from the shared template, not the same underlying request." + }, + { + "left": "issue:27088", + "right": "issue:44365", + "accept": false, + "reason": "Turkish-specific request vs a generic language-template issue; no evidence they target the same language or same translation task." + }, + { + "left": "issue:27088", + "right": "issue:43834", + "accept": false, + "reason": "Turkish-specific request vs a generic language-template issue; similar wording alone is not enough to treat them as the same issue." + }, + { + "left": "issue:43834", + "right": "issue:44370", + "accept": false, + "reason": "Both appear to be template-style translation requests, but there is no concrete evidence they are the same language or the same underlying task." + }, + { + "left": "issue:43834", + "right": "issue:44365", + "accept": false, + "reason": "Identical placeholder-style titles are not enough to merge; they may correspond to different language-specific requests." + }, + { + "left": "issue:44365", + "right": "issue:44370", + "accept": false, + "reason": "Both are generic translation-request templates, but no explicit language target or shared concrete bug/change is shown." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 5, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-36817-15", + "nodes": [ + "issue:36817", + "issue:36879", + "issue:37015", + "issue:41115", + "issue:41211", + "issue:42503", + "issue:42738", + "issue:43671", + "issue:43837", + "issue:43909", + "issue:44121", + "issue:44205", + "issue:45027", + "issue:45177", + "issue:45322" + ], + "soft_pairs": [ + "issue:36817|issue:41211", + "issue:36817|issue:43671", + "issue:36817|issue:36879", + "issue:36879|issue:41211", + "issue:36879|issue:43909", + "issue:36879|issue:44121", + "issue:36879|issue:43837", + "issue:36879|issue:43671", + "issue:37015|issue:43837", + "issue:37015|issue:44121", + "issue:37015|issue:43909", + "issue:36879|issue:37015", + "issue:37015|issue:42503", + "issue:41115|issue:43837", + "issue:41115|issue:44121", + "issue:41115|issue:42503", + "issue:41115|issue:45027", + "issue:41211|issue:43671", + "issue:42503|issue:43837", + "issue:42503|issue:43671", + "issue:42503|issue:42738", + "issue:42738|issue:45027", + "issue:43671|issue:43837", + "issue:43671|issue:45027", + "issue:43671|issue:45177", + "issue:43671|issue:44121", + "issue:43671|issue:43909", + "issue:43837|issue:43909", + "issue:43837|issue:45322", + "issue:43837|issue:44121", + "issue:43909|issue:44205", + "issue:44121|issue:45322", + "issue:43837|issue:45027", + "issue:37015|issue:45027", + "issue:43837|issue:45177" + ], + "prepared_review_unit_hash": "ea29828d0bdc27ecc969455f3ab60d3576f0860d8e95d5078852e49cbba68b75", + "component_budget": { + "node_count": 15, + "item_count": 15, + "soft_pair_count": 35, + "serialized_chars": 20712, + "estimated_input_tokens": 5178, + "estimated_eval_tokens": 10612 + }, + "budget": { + "node_count": 15, + "item_count": 15, + "soft_pair_count": 35, + "serialized_chars": 20712, + "estimated_input_tokens": 5178, + "estimated_eval_tokens": 10612 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "This cluster is a loose theme cluster of model-add/support requests, not true duplicates. The issues name different models or different modalities, so they should stay separate; at most, a couple are nearby subfamily requests.", + "confidence": 0.84, + "canonical_issue_reason": "If one issue must act as the anchor, issue 43837 is the best fit because it sits in the most connected subcluster and has the strongest overlap with a nearby related request (43671), but the set is still not duplicate enough to collapse into one issue.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 43671 is the best representative of the cluster because it is a clear, specific model-support request with substantial discussion and belongs to the densest nearby subcluster; however, it still represents a distinct request from the others.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:36817", + "right": "issue:41211", + "accept": false, + "reason": "Different model requests (EuroBert vs DEIMv2); same broad pattern, not the same bug/change." + }, + { + "left": "issue:36817", + "right": "issue:43671", + "accept": false, + "reason": "Different requested models and different modality area; not a duplicate." + }, + { + "left": "issue:36817", + "right": "issue:36879", + "accept": false, + "reason": "Both are model-add requests, but for unrelated models and code paths." + }, + { + "left": "issue:36879", + "right": "issue:41211", + "accept": false, + "reason": "RF-DETR and DEIMv2 are distinct model integrations." + }, + { + "left": "issue:36879", + "right": "issue:43909", + "accept": false, + "reason": "Detection model request vs audio model request; not the same change." + }, + { + "left": "issue:36879", + "right": "issue:44121", + "accept": false, + "reason": "Different model families and integration work; no single concrete fix." + }, + { + "left": "issue:36879", + "right": "issue:43837", + "accept": false, + "reason": "Different requested models; same support-request theme only." + }, + { + "left": "issue:36879", + "right": "issue:43671", + "accept": false, + "reason": "RF-DETR vs Qwen3-TTS are unrelated model additions." + }, + { + "left": "issue:37015", + "right": "issue:43837", + "accept": false, + "reason": "NeoBERT and Qwen3-ASR are separate model support requests." + }, + { + "left": "issue:37015", + "right": "issue:44121", + "accept": false, + "reason": "Different models with different implementation paths." + }, + { + "left": "issue:37015", + "right": "issue:43909", + "accept": false, + "reason": "NeoBERT and LFM2.5 Audio are unrelated requests." + }, + { + "left": "issue:36879", + "right": "issue:37015", + "accept": false, + "reason": "Different models; only the generic request template matches." + }, + { + "left": "issue:37015", + "right": "issue:42503", + "accept": false, + "reason": "NeoBERT and ModernVBERT are distinct model requests, not duplicates." + }, + { + "left": "issue:41115", + "right": "issue:43837", + "accept": false, + "reason": "MiniCPM3 architecture request vs Qwen3-ASR support; different artifacts." + }, + { + "left": "issue:41115", + "right": "issue:44121", + "accept": false, + "reason": "Different model families and likely different files/code paths." + }, + { + "left": "issue:41115", + "right": "issue:42503", + "accept": false, + "reason": "MiniCPM3 and ModernVBERT are separate model additions." + }, + { + "left": "issue:41115", + "right": "issue:45027", + "accept": false, + "reason": "MiniCPM3 architecture vs Voxtral TTS support; not the same change." + }, + { + "left": "issue:41211", + "right": "issue:43671", + "accept": false, + "reason": "DEIMv2 and Qwen3-TTS are unrelated model requests." + }, + { + "left": "issue:42503", + "right": "issue:43837", + "accept": false, + "reason": "ModernVBERT and Qwen3-ASR are different models." + }, + { + "left": "issue:42503", + "right": "issue:43671", + "accept": false, + "reason": "ModernVBERT models vs Qwen3-TTS support; not a duplicate." + }, + { + "left": "issue:42503", + "right": "issue:42738", + "accept": false, + "reason": "ModernVBERT and BERT-like models with RoPE are different requests; too broad to merge." + }, + { + "left": "issue:42738", + "right": "issue:45027", + "accept": false, + "reason": "A generic BERT-like/RoPE request is not the same as Voxtral-4B-TTS support." + }, + { + "left": "issue:43671", + "right": "issue:43837", + "accept": false, + "reason": "Same Qwen3 family, but TTS vs ASR are different products and code paths." + }, + { + "left": "issue:43671", + "right": "issue:45027", + "accept": false, + "reason": "Both are speech-related, but they request different model support." + }, + { + "left": "issue:43671", + "right": "issue:45177", + "accept": false, + "reason": "Qwen3-TTS and DeepSeek-OCR2 are unrelated model additions." + }, + { + "left": "issue:43671", + "right": "issue:44121", + "accept": false, + "reason": "Different model families and different integration work." + }, + { + "left": "issue:43671", + "right": "issue:43909", + "accept": false, + "reason": "Qwen3-TTS and LFM2.5 Audio are distinct audio-model requests." + }, + { + "left": "issue:43837", + "right": "issue:43909", + "accept": false, + "reason": "Qwen3-ASR and LFM2.5 Audio are not the same model or fix." + }, + { + "left": "issue:43837", + "right": "issue:45322", + "accept": false, + "reason": "Qwen3-ASR support and EUPE are unrelated model requests." + }, + { + "left": "issue:43837", + "right": "issue:44121", + "accept": false, + "reason": "Qwen3-ASR and the weight-sparse transformer are distinct requests." + }, + { + "left": "issue:43909", + "right": "issue:44205", + "accept": false, + "reason": "Audio model support vs SAM3-LiteText; different model integrations." + }, + { + "left": "issue:44121", + "right": "issue:45322", + "accept": false, + "reason": "Weight-sparse transformer and EUPE are separate model requests." + }, + { + "left": "issue:43837", + "right": "issue:45027", + "accept": false, + "reason": "Both are speech-related, but ASR and TTS support are different concrete changes." + }, + { + "left": "issue:37015", + "right": "issue:45027", + "accept": false, + "reason": "NeoBERT and Voxtral-4B-TTS are unrelated model additions." + }, + { + "left": "issue:43837", + "right": "issue:45177", + "accept": false, + "reason": "Qwen3-ASR and DeepSeek-OCR2 do not target the same model or code path." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 6, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-37934-6", + "nodes": [ + "issue:37934", + "issue:42936", + "issue:43240", + "issue:43668", + "issue:43813", + "issue:44031" + ], + "soft_pairs": [ + "issue:37934|issue:43668", + "issue:42936|issue:43240", + "issue:42936|issue:44031", + "issue:42936|issue:43813", + "issue:42936|issue:43668", + "issue:43240|issue:43668", + "issue:43668|issue:44031", + "issue:43668|issue:43813" + ], + "prepared_review_unit_hash": "6f142b0557b4ef80729795269016402f28030efc2717b63aa38ccea90564d478", + "component_budget": { + "node_count": 6, + "item_count": 6, + "soft_pair_count": 8, + "serialized_chars": 6291, + "estimated_input_tokens": 1573, + "estimated_eval_tokens": 3402 + }, + "budget": { + "node_count": 6, + "item_count": 6, + "soft_pair_count": 8, + "serialized_chars": 6291, + "estimated_input_tokens": 1573, + "estimated_eval_tokens": 3402 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These issues are only loosely text-similar and describe different bugs across unrelated areas (model norms, config typing, loss kwargs, a typo, tokenizer warnings). I do not see a true duplicate cluster here.", + "confidence": 0.96, + "canonical_issue_reason": "No single canonical issue: the items cover unrelated problem types and code paths, so none is a good duplicate anchor.", + "canonical_pr_reason": null, + "best_issue_reason": "No best issue for deduplication because the set is heterogeneous rather than one underlying bug.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:37934", + "right": "issue:43668", + "accept": false, + "reason": "Both mention normalization/config names, but one questions an architecture choice in Llama4 while the other is a type-hint issue in ModernBERTConfig; not the same bug." + }, + { + "left": "issue:42936", + "right": "issue:43240", + "accept": false, + "reason": "Mask2Former ignore_value handling and loss kwargs plumbing are different parts of the training stack and different failure modes." + }, + { + "left": "issue:42936", + "right": "issue:44031", + "accept": false, + "reason": "Unrelated topics: a model loss parameter not used versus tokenizer regex warning regression." + }, + { + "left": "issue:42936", + "right": "issue:43813", + "accept": false, + "reason": "A Mask2Former loss-parameter issue is unrelated to a typo in a quantization-operation attribute name." + }, + { + "left": "issue:42936", + "right": "issue:43668", + "accept": false, + "reason": "Different models and different bugs: ignore_value usage versus an incorrect type hint." + }, + { + "left": "issue:43240", + "right": "issue:43668", + "accept": false, + "reason": "Loss-kwargs propagation and ModernBERTConfig type annotation are separate issues with no shared code path." + }, + { + "left": "issue:43668", + "right": "issue:44031", + "accept": false, + "reason": "A config type hint bug does not match a tokenizer-regex warning regression." + }, + { + "left": "issue:43668", + "right": "issue:43813", + "accept": false, + "reason": "ModernBERTConfig typing and an attribute-name typo are distinct problems in different code areas." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 7, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-41084-2", + "nodes": [ + "issue:41084", + "issue:43037" + ], + "soft_pairs": [ + "issue:41084|issue:43037" + ], + "prepared_review_unit_hash": "f2f820abf20784a482a6536760011db0bb596b7a6121c03cc6d888fbfaf5586e", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1478, + "estimated_input_tokens": 370, + "estimated_eval_tokens": 996 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1478, + "estimated_input_tokens": 370, + "estimated_eval_tokens": 996 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The two issues are not duplicates: one is a vague feature request for \"Set Block Decoding,\" while the other is a specific request for \"DeepSeek v3.2 support.\" They appear to be unrelated product/runtime feature requests rather than the same underlying problem.", + "confidence": 0.92, + "canonical_issue_reason": "Issue #43037 is the better canonical anchor because it is more specific and actionable; #41084 is too vague to serve as a stable representative for the cluster.", + "canonical_pr_reason": null, + "best_issue_reason": "#43037 is the best standalone issue candidate due to its concrete scope and clearer target. #41084 is too underspecified to be the primary reference issue.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:41084", + "right": "issue:43037", + "accept": false, + "reason": "The titles and available metadata point to different requests: block decoding behavior versus DeepSeek v3.2 model support. No shared files, targets, or issue links suggest the same underlying change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 8, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43598-2", + "nodes": [ + "issue:43598", + "issue:43600" + ], + "soft_pairs": [ + "issue:43598|issue:43600" + ], + "prepared_review_unit_hash": "ea4e7f13aec445d58bcdb636eb9f774f112fcc06e0aa978d7f87c8862aa99c46", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1584, + "estimated_input_tokens": 396, + "estimated_eval_tokens": 1048 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1584, + "estimated_input_tokens": 396, + "estimated_eval_tokens": 1048 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These two issues are about different trainer behaviors: one concerns `remove_unused_column` customizability, the other concerns FLOPS estimation for custom trainers. They do not appear to describe the same underlying bug or change.", + "confidence": 0.96, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43598", + "right": "issue:43600", + "accept": false, + "reason": "Different problem domains and code paths: one is about column pruning customization in Trainer, the other about FLOPS accounting for custom trainers. Shared theme is only broad Trainer customization, not a duplicate bug/change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 9, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43684-2", + "nodes": [ + "issue:43684", + "issue:44405" + ], + "soft_pairs": [ + "issue:43684|issue:44405" + ], + "prepared_review_unit_hash": "c42d55c30f6c06de9da09976bc911fd47df1406b9fad1a8084703601c9c1e033", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1612, + "estimated_input_tokens": 403, + "estimated_eval_tokens": 1062 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1612, + "estimated_input_tokens": 403, + "estimated_eval_tokens": 1062 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both issues are about extending Hugging Face auto-class support for Qwen-family models, but they request different changes: Qwen3-Omni registration in core auto classes vs. Qwen3.5 sequence-classification support. They are related feature requests, not the same underlying bug/change.", + "confidence": 0.9, + "canonical_issue_reason": "issue:43684 is the broader auto-registration request and best represents the cluster\u2019s general theme, but it is still not the same as the Qwen3.5 sequence-classification request.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43684 is the better cluster representative because it targets core AutoModel/AutoModelForConditionalGeneration registration rather than a narrower task-specific head.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43684", + "right": "issue:44405", + "accept": false, + "reason": "Different model variants and different auto-class targets; related in theme, but not the same concrete issue or change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 10, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43835-4", + "nodes": [ + "issue:43835", + "issue:43845", + "issue:43846", + "issue:43859" + ], + "soft_pairs": [ + "issue:43835|issue:43845", + "issue:43835|issue:43846", + "issue:43835|issue:43859", + "issue:43845|issue:43846", + "issue:43845|issue:43859", + "issue:43846|issue:43859" + ], + "prepared_review_unit_hash": "46073e5cddb5c8499c487ae96083e3eca4e454d69aaeae59b362c4f3cfc03ee8", + "component_budget": { + "node_count": 4, + "item_count": 4, + "soft_pair_count": 6, + "serialized_chars": 4258, + "estimated_input_tokens": 1065, + "estimated_eval_tokens": 2386 + }, + "budget": { + "node_count": 4, + "item_count": 4, + "soft_pair_count": 6, + "serialized_chars": 4258, + "estimated_input_tokens": 1065, + "estimated_eval_tokens": 2386 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "#43845 is the best canonical issue among the three near-identical \"huggingface \" reports because it was opened first. #43835 looks unrelated. The similarity between the huggingface issues is too shallow to confidently mark any pair as duplicates without body details or explicit links.", + "confidence": 0.78, + "canonical_issue_reason": "#43845 is the earliest of the three near-identical \"huggingface \" issues, so it is the best representative of that cluster.", + "canonical_pr_reason": null, + "best_issue_reason": "#43845 is the strongest issue candidate because it is the first report in the placeholder-title cluster and is the most natural canonical representative.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43835", + "right": "issue:43845", + "accept": false, + "reason": "Different titles and no concrete shared bug context; the evidence is only weak text similarity." + }, + { + "left": "issue:43835", + "right": "issue:43846", + "accept": false, + "reason": "Different titles and no shared details point to unrelated reports rather than the same issue." + }, + { + "left": "issue:43835", + "right": "issue:43859", + "accept": false, + "reason": "\"Fett\" does not plausibly describe the same underlying problem as the huggingface issues." + }, + { + "left": "issue:43845", + "right": "issue:43846", + "accept": false, + "reason": "Same placeholder title, but only soft similarity is available and deterministic acceptance is false, so this is not enough to confirm a duplicate." + }, + { + "left": "issue:43845", + "right": "issue:43859", + "accept": false, + "reason": "The titles are similar, but there are no body details or explicit links tying them to the same underlying bug." + }, + { + "left": "issue:43846", + "right": "issue:43859", + "accept": false, + "reason": "These look superficially alike, but the available evidence is too weak to treat them as duplicates with confidence." + } + ] + }, + "evaluator_result": { + "accept": false, + "feedback": "The canonical issue choice is reasonably grounded: #43845 is the earliest of the three issues with the same title. However, the soft-edge verdicts are too strong for the packet: all pairings are supported only by soft_similarity, and the packet\u2019s deterministic_accept is false for every pair. With no issue-body details or explicit links, it is not conservative to mark the #43845/#43846/#43859 pairs as duplicates with high confidence. Rejecting the #43835 comparisons is fine, but the duplicate conclusions among the huggingface issues overstate the evidence." + }, + "evaluator_used": true, + "retried": true, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 11, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-44371-5", + "nodes": [ + "issue:44371", + "issue:44418", + "issue:44572", + "issue:44573", + "issue:44937" + ], + "soft_pairs": [ + "issue:44371|issue:44572", + "issue:44371|issue:44573", + "issue:44371|issue:44418", + "issue:44371|issue:44937", + "issue:44418|issue:44572", + "issue:44418|issue:44573", + "issue:44572|issue:44573", + "issue:44572|issue:44937", + "issue:44573|issue:44937" + ], + "prepared_review_unit_hash": "5240be50cddc1c18171a3108f04aac3f19fa353feaeeb470bc99c61a7ce9c13e", + "component_budget": { + "node_count": 5, + "item_count": 5, + "soft_pair_count": 9, + "serialized_chars": 5791, + "estimated_input_tokens": 1448, + "estimated_eval_tokens": 3152 + }, + "budget": { + "node_count": 5, + "item_count": 5, + "soft_pair_count": 9, + "serialized_chars": 5791, + "estimated_input_tokens": 1448, + "estimated_eval_tokens": 3152 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are not clear duplicates of one bug/feature; they look like unrelated low-signal/spam issues with only superficial textual similarity. No soft pair should be merged.", + "confidence": 0.87, + "canonical_issue_reason": "issue:44418 is the least spam-like and most issue-shaped title in the set, so it is the best forced representative if one must be chosen.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:44418 is the most plausible standalone issue title here; the others are generic spam placeholders or unrelated text.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44371", + "right": "issue:44572", + "accept": false, + "reason": "Both are just '' with no concrete shared bug or change; no evidence they describe the same underlying issue." + }, + { + "left": "issue:44371", + "right": "issue:44573", + "accept": false, + "reason": "Both are spam placeholders and provide no shared problem description beyond the title text." + }, + { + "left": "issue:44371", + "right": "issue:44418", + "accept": false, + "reason": "One is spam and the other is a generic documentation suggestion; not the same underlying issue." + }, + { + "left": "issue:44371", + "right": "issue:44937", + "accept": false, + "reason": "'' and 'Check out \"Google LLC\"' are unrelated titles with no concrete bug overlap." + }, + { + "left": "issue:44418", + "right": "issue:44572", + "accept": false, + "reason": "Documentation enhancement request does not match a spam placeholder issue; no shared problem statement." + }, + { + "left": "issue:44418", + "right": "issue:44573", + "accept": false, + "reason": "Generic documentation request and spam placeholder are unrelated." + }, + { + "left": "issue:44572", + "right": "issue:44573", + "accept": false, + "reason": "Same spam marker, but that alone is not enough to indicate the same underlying bug or change." + }, + { + "left": "issue:44572", + "right": "issue:44937", + "accept": false, + "reason": "Spam placeholder and unrelated Google LLC title do not point to the same issue." + }, + { + "left": "issue:44573", + "right": "issue:44937", + "accept": false, + "reason": "Spam placeholder and unrelated Google LLC title do not describe the same bug or request." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 12, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45145-2", + "nodes": [ + "issue:45145", + "issue:45151" + ], + "soft_pairs": [ + "issue:45145|issue:45151" + ], + "prepared_review_unit_hash": "44f6144620f78e303387c8dab336fc1e19178192af2e148f87351790478c33c9", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1645, + "estimated_input_tokens": 412, + "estimated_eval_tokens": 1080 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1645, + "estimated_input_tokens": 412, + "estimated_eval_tokens": 1080 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The two issues are highly similar and may be duplicates: they share the exact title, have very close body lengths, and there are no distinguishing links, targets, or filenames. However, the packet does not prove they are exact duplicates, so the safer conclusion is likely-duplicate rather than confirmed duplicate. #45145 is the better canonical record because it was created first and has an inbound reference.", + "confidence": 0.83, + "canonical_issue_reason": "#45145 is the best canonical issue because it is earlier and has an inbound reference. The evidence supports strong similarity with #45151, but not a definitive exact-duplicate determination.", + "canonical_pr_reason": null, + "best_issue_reason": "#45145 is the best representative issue for the cluster: it is the earlier report and has slightly more established history via the inbound reference, while #45151 does not add distinct scope.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45145", + "right": "issue:45151", + "accept": false, + "reason": "The pair looks highly similar, but the evidence is only soft similarity plus matching titles. With no explicit links, targets, or content proof, this is not enough to confidently accept them as the same underlying issue." + } + ] + }, + "evaluator_result": { + "accept": false, + "feedback": "The canonical choice is plausible, but the writeup overstates the evidence. The packet supports that the two issues have the same title and very similar metadata/body lengths, with no explicit links or targets, but it does not establish that they are exact duplicates or that the issue content is the same. The soft pair is also only a similarity signal and the deterministic decision is false, so the duplicate verdict and 0.99 confidence are too strong. A more conservative summary would say they appear highly similar / likely duplicates, with #45145 as the better canonical record because it is earlier and has an inbound reference." + }, + "evaluator_used": true, + "retried": true, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 13, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45295-2", + "nodes": [ + "issue:45295", + "issue:45373" + ], + "soft_pairs": [], + "prepared_review_unit_hash": "f5fd95fada2f530ab0999e343c17c16266cde0c7182bb22509972ca44827cd49", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 0, + "serialized_chars": 1286, + "estimated_input_tokens": 322, + "estimated_eval_tokens": 900 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 0, + "serialized_chars": 1286, + "estimated_input_tokens": 322, + "estimated_eval_tokens": 900 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These two issues describe the same missing Gemma 4 sequence-classification support; the later, more specific open issue is the better cluster representative.", + "confidence": 0.94, + "canonical_issue_reason": "issue #45373 is the clearest canonical tracker: it is open, explicitly names the missing class `Gemma4ForSequenceClassification`, and states the module-level gap relative to Gemma 2/3.", + "canonical_pr_reason": null, + "best_issue_reason": "issue #45373 best captures the concrete request and likely remaining work, while #45295 is a shorter earlier duplicate with less detail and is already closed.", + "best_pr_reason": null, + "soft_edge_verdicts": [] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 14, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45672-2", + "nodes": [ + "issue:45672", + "issue:45674" + ], + "soft_pairs": [ + "issue:45672|issue:45674" + ], + "prepared_review_unit_hash": "65b21d096078d27c4cceb409f978dd81a76bcc6de62d377dfa17d102c8c99c21", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1668, + "estimated_input_tokens": 417, + "estimated_eval_tokens": 1090 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1668, + "estimated_input_tokens": 417, + "estimated_eval_tokens": 1090 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These two issues both mention 4-bit inference, but they describe different failure modes and likely different code paths: one is a Gemma4/audio-module uint8 dtype error, the other is a BitsAndBytesConfig skip-module regression that removes the default lm_head exclusion. I would keep them separate and not merge the cluster.", + "confidence": 0.91, + "canonical_issue_reason": "Issue #45674 is the better cluster representative because it describes a concrete, broadly applicable BitsAndBytesConfig regression with a clear expected/actual behavior and an obvious configuration-level fix path.", + "canonical_pr_reason": null, + "best_issue_reason": "#45674 is the strongest issue to anchor on: it is specific, reproducible, and appears to describe the main actionable bug in the cluster. #45672 is a narrower model-specific dtype crash in audio modules and does not match the same root cause.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45672", + "right": "issue:45674", + "accept": false, + "reason": "Both are 4-bit inference bugs, but the failure mechanisms differ: #45672 is a Gemma4/audio uint8 torch.finfo TypeError, while #45674 is a BitsAndBytesConfig skip-module bug causing lm_head exclusion to be cleared and an AssertionError. They do not look like the same underlying issue." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 15, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45762-2", + "nodes": [ + "issue:45762", + "issue:45763" + ], + "soft_pairs": [ + "issue:45762|issue:45763" + ], + "prepared_review_unit_hash": "7d1384808e2e81500921827e3a7a187a03629af2c2f7ed4f132e07478de6c054", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1477, + "estimated_input_tokens": 370, + "estimated_eval_tokens": 996 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1477, + "estimated_input_tokens": 370, + "estimated_eval_tokens": 996 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both items look like unrelated low-information issues/spam entries (personal names plus birth dates), not duplicate bug reports. There is no meaningful canonical issue or PR here.", + "confidence": 0.97, + "canonical_issue_reason": "None: neither issue contains a bug report, repro, or substantive discussion to serve as a canonical duplicate target.", + "canonical_pr_reason": null, + "best_issue_reason": "Neither issue is suitable as a global best issue; both are empty placeholder-like entries with no actionable content.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45762", + "right": "issue:45763", + "accept": false, + "reason": "The titles are just different personal names and dates; there is no shared bug, feature, or code-path problem to indicate duplication." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 16, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45764-2", + "nodes": [ + "issue:45764", + "issue:45765" + ], + "soft_pairs": [ + "issue:45764|issue:45765" + ], + "prepared_review_unit_hash": "110af76adcf80338b6b3aa8304f9685c59f559825f900f0c9c81faedbffdeb5a", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1481, + "estimated_input_tokens": 371, + "estimated_eval_tokens": 998 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1481, + "estimated_input_tokens": 371, + "estimated_eval_tokens": 998 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The two issues are just separate name-and-date titles with no body, links, or shared context; there is no evidence they describe the same bug or request.", + "confidence": 0.95, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45764", + "right": "issue:45765", + "accept": false, + "reason": "No substantive overlap beyond both being bare personal-name titles; they do not appear to report the same underlying problem." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 17, + "component_count": 17, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45869-5", + "nodes": [ + "issue:45869", + "issue:45870", + "issue:45871", + "issue:45872", + "issue:45873" + ], + "soft_pairs": [ + "issue:45869|issue:45870", + "issue:45869|issue:45871", + "issue:45870|issue:45871", + "issue:45870|issue:45872", + "issue:45870|issue:45873", + "issue:45871|issue:45872", + "issue:45871|issue:45873", + "issue:45872|issue:45873" + ], + "prepared_review_unit_hash": "d44579cb77a0837216217ddc5acd11ab1afa90d63ab85500a28997e9dc6d6dee", + "component_budget": { + "node_count": 5, + "item_count": 5, + "soft_pair_count": 8, + "serialized_chars": 5726, + "estimated_input_tokens": 1432, + "estimated_eval_tokens": 3120 + }, + "budget": { + "node_count": 5, + "item_count": 5, + "soft_pair_count": 8, + "serialized_chars": 5726, + "estimated_input_tokens": 1432, + "estimated_eval_tokens": 3120 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Two exact duplicate issue pairs are present: 45870/45871 and 45872/45873. Issue 45869 appears to be a separate dataset/sample request and does not match the others.", + "confidence": 0.97, + "canonical_issue_reason": "45870 is the earliest of the exact duplicate pair 45870/45871 and has the cleanest representative title for that topic.", + "canonical_pr_reason": null, + "best_issue_reason": "45870 is the best single issue to keep for its duplicate pair because it is the earliest, clear, and exact match to 45871.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45869", + "right": "issue:45870", + "accept": false, + "reason": "Different requests: a financial dataset bundle versus a DataForge prediction example." + }, + { + "left": "issue:45869", + "right": "issue:45871", + "accept": false, + "reason": "Different requests: a financial dataset bundle versus a DataForge prediction example." + }, + { + "left": "issue:45870", + "right": "issue:45871", + "accept": true, + "reason": "Exact duplicate titles and same described example request." + }, + { + "left": "issue:45870", + "right": "issue:45872", + "accept": false, + "reason": "Different topics: DataForge prediction example versus DataForge plus scraping service." + }, + { + "left": "issue:45870", + "right": "issue:45873", + "accept": false, + "reason": "Different topics: DataForge prediction example versus DataForge plus scraping service." + }, + { + "left": "issue:45871", + "right": "issue:45872", + "accept": false, + "reason": "Different topics: DataForge prediction example versus DataForge plus scraping service." + }, + { + "left": "issue:45871", + "right": "issue:45873", + "accept": false, + "reason": "Different topics: DataForge prediction example versus DataForge plus scraping service." + }, + { + "left": "issue:45872", + "right": "issue:45873", + "accept": true, + "reason": "Exact duplicate titles and same described DataForge scraping-service request." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "Grounded overall. The summary conservatively separates 45869 from the two duplicate-looking pairs, and the pair verdicts match the packet: 45870/45871 and 45872/45873 have identical titles and jaccard 1.0, while the cross-topic pairs are reasonably rejected." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 1, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-39895-16", + "nodes": [ + "pull_request:39895", + "pull_request:41855", + "pull_request:42668", + "pull_request:43251", + "pull_request:43254", + "pull_request:43424", + "pull_request:43823", + "pull_request:43995", + "pull_request:44827", + "pull_request:45170", + "pull_request:45213", + "pull_request:45655", + "pull_request:45795", + "pull_request:45899", + "pull_request:45919", + "pull_request:45922" + ], + "soft_pairs": [ + "pull_request:39895|pull_request:43424", + "pull_request:41855|pull_request:43823", + "pull_request:41855|pull_request:43424", + "pull_request:42668|pull_request:43823", + "pull_request:43251|pull_request:43823", + "pull_request:43251|pull_request:44827", + "pull_request:43424|pull_request:45919", + "pull_request:43424|pull_request:45899", + "pull_request:43424|pull_request:44827", + "pull_request:43424|pull_request:43823", + "pull_request:43823|pull_request:45655", + "pull_request:43823|pull_request:45922", + "pull_request:43823|pull_request:43995", + "pull_request:43995|pull_request:45213", + "pull_request:43995|pull_request:44827", + "pull_request:43995|pull_request:45899", + "pull_request:44827|pull_request:45899", + "pull_request:44827|pull_request:45170", + "pull_request:44827|pull_request:45795", + "pull_request:44827|pull_request:45655", + "pull_request:45170|pull_request:45899", + "pull_request:44827|pull_request:45213", + "pull_request:43823|pull_request:45213", + "pull_request:45213|pull_request:45899", + "pull_request:45795|pull_request:45899", + "pull_request:43823|pull_request:45795", + "pull_request:43823|pull_request:45899", + "pull_request:44827|pull_request:45922", + "pull_request:43251|pull_request:43254" + ], + "prepared_review_unit_hash": "b55d20d144ef80c3c13b8376a5e390e335227a9a88d87f7488b18563914f97af", + "component_budget": { + "node_count": 16, + "item_count": 16, + "soft_pair_count": 29, + "serialized_chars": 46702, + "estimated_input_tokens": 11676, + "estimated_eval_tokens": 23608 + }, + "budget": { + "node_count": 16, + "item_count": 16, + "soft_pair_count": 29, + "serialized_chars": 46702, + "estimated_input_tokens": 11676, + "estimated_eval_tokens": 23608 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "This cluster is mostly a grab bag of unrelated PRs (new model additions, docs, and refactors). The only true duplicate pair is the two `fixed_cross_entropy` changes; the rest should stay separate.", + "confidence": 0.93, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 43251 is the best canonical PR for the only real duplicate pair: it directly targets issue 43240 and contains the fuller `fixed_cross_entropy` kwargs-forwarding change, while 43254 is a narrower variant.", + "best_issue_reason": null, + "best_pr_reason": "PR 43251 is the strongest representative of the actual duplicate change because it is the more complete implementation of the same cross-entropy fix and has the explicit issue linkage; the other PRs are not duplicates of it.", + "soft_edge_verdicts": [ + { + "left": "pull_request:39895", + "right": "pull_request:43424", + "accept": false, + "reason": "Videoprism model addition vs ExecuTorch dynamic-shape test; different subsystems and no shared bug/change." + }, + { + "left": "pull_request:41855", + "right": "pull_request:43823", + "accept": false, + "reason": "Mistral tokenizer API additions vs adding a new MobileLLM model; unrelated changes." + }, + { + "left": "pull_request:41855", + "right": "pull_request:43424", + "accept": false, + "reason": "Tokenizer utility fix vs export test coverage; not the same code-path problem." + }, + { + "left": "pull_request:42668", + "right": "pull_request:43823", + "accept": false, + "reason": "Processor auto-mapping robustness work vs new MobileLLM model implementation; different underlying changes." + }, + { + "left": "pull_request:43251", + "right": "pull_request:43823", + "accept": false, + "reason": "Cross-entropy helper fix vs new model addition; no overlap beyond being Transformers PRs." + }, + { + "left": "pull_request:43251", + "right": "pull_request:44827", + "accept": false, + "reason": "Loss-function kwargs fix and Mistral4 test/model work are unrelated." + }, + { + "left": "pull_request:43424", + "right": "pull_request:45919", + "accept": false, + "reason": "ExecuTorch export testing vs adding Sapiens2; different goals and code paths." + }, + { + "left": "pull_request:43424", + "right": "pull_request:45899", + "accept": false, + "reason": "Export dynamic-shape test vs docs-only continuous batching note; not a duplicate change." + }, + { + "left": "pull_request:43424", + "right": "pull_request:44827", + "accept": false, + "reason": "ExecuTorch export test and Mistral4 test/model fixes are unrelated." + }, + { + "left": "pull_request:43424", + "right": "pull_request:43823", + "accept": false, + "reason": "Export testing vs MobileLLM model addition; not the same underlying bug/change." + }, + { + "left": "pull_request:43823", + "right": "pull_request:45655", + "accept": false, + "reason": "New MobileLLM model vs config-resolution refactor; unrelated." + }, + { + "left": "pull_request:43823", + "right": "pull_request:45922", + "accept": false, + "reason": "New MobileLLM model vs vision-model lru cache leak fix; different code paths." + }, + { + "left": "pull_request:43823", + "right": "pull_request:43995", + "accept": false, + "reason": "MobileLLM model addition vs Falcon output-interface refactor; no duplicate relationship." + }, + { + "left": "pull_request:43995", + "right": "pull_request:45213", + "accept": false, + "reason": "Falcon model refactor vs model-creation skill/docs; completely different changes." + }, + { + "left": "pull_request:43995", + "right": "pull_request:44827", + "accept": false, + "reason": "Falcon output-collection refactor vs Mistral4 test fixes; not the same fix." + }, + { + "left": "pull_request:43995", + "right": "pull_request:45899", + "accept": false, + "reason": "Falcon model refactor vs continuous-batching docs update; unrelated." + }, + { + "left": "pull_request:44827", + "right": "pull_request:45899", + "accept": false, + "reason": "Mistral4 test/model fixes vs docs for decode fast path; not the same underlying change." + }, + { + "left": "pull_request:44827", + "right": "pull_request:45170", + "accept": false, + "reason": "Mistral4 test fixes vs `layrnorm` renaming pass; different models and problems." + }, + { + "left": "pull_request:44827", + "right": "pull_request:45795", + "accept": false, + "reason": "Mistral4 test fixes vs docs for audio/video processors; unrelated." + }, + { + "left": "pull_request:44827", + "right": "pull_request:45655", + "accept": false, + "reason": "Mistral4 fix vs config-resolution order fix; different issues." + }, + { + "left": "pull_request:45170", + "right": "pull_request:45899", + "accept": false, + "reason": "CLIP-like weight renaming vs docs-only batching writeup; no shared code path." + }, + { + "left": "pull_request:44827", + "right": "pull_request:45213", + "accept": false, + "reason": "Shared auto/modeling_auto filename is incidental; Mistral4 tests and model-creation skill are unrelated." + }, + { + "left": "pull_request:43823", + "right": "pull_request:45213", + "accept": false, + "reason": "New model addition vs skill/doc scaffolding for model creation; not the same change." + }, + { + "left": "pull_request:45213", + "right": "pull_request:45899", + "accept": false, + "reason": "Model-creation skill PR vs continuous-batching docs; unrelated documentation areas." + }, + { + "left": "pull_request:45795", + "right": "pull_request:45899", + "accept": false, + "reason": "Docs for audio/video processors vs docs for decode fast path; different documentation topics." + }, + { + "left": "pull_request:43823", + "right": "pull_request:45795", + "accept": false, + "reason": "MobileLLM model addition vs processor docs; no duplicate underlying bug/change." + }, + { + "left": "pull_request:43823", + "right": "pull_request:45899", + "accept": false, + "reason": "New model addition vs continuous-batching docs update; unrelated." + }, + { + "left": "pull_request:44827", + "right": "pull_request:45922", + "accept": false, + "reason": "Mistral4 test fixes vs vision-model memory-leak fix; different model families and defects." + }, + { + "left": "pull_request:43251", + "right": "pull_request:43254", + "accept": true, + "reason": "Both PRs patch `fixed_cross_entropy` in the same file for the same issue (43240) and differ only in how the supported kwargs are threaded through; they could be merged into one fix." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "Grounded and conservative overall. The only pair that looks like a real duplicate bugfix is PR 43251/43254: both edit `src/transformers/loss/loss_utils.py` for issue 43240 and are mergeable into one `fixed_cross_entropy` kwargs-forwarding fix. The other pair verdicts are appropriately negative given the very different subsystems and file sets, even where the similarity heuristic was noisy." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 2, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-40546-2", + "nodes": [ + "pull_request:40546", + "pull_request:42781" + ], + "soft_pairs": [ + "pull_request:40546|pull_request:42781" + ], + "prepared_review_unit_hash": "56d44b2db28df2867e1839c908b5c8418f06309ef086e934b11c8aa32c1edb0b", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 7226, + "estimated_input_tokens": 1807, + "estimated_eval_tokens": 3870 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 7226, + "estimated_input_tokens": 1807, + "estimated_eval_tokens": 3870 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Two related but distinct VibeVoice PRs: one implements the base VibeVoice integration, the other adds VibeVoice Realtime with extra acoustic-tokenizer and generation/model changes. They overlap on shared package wiring, but they are not the same underlying change.", + "confidence": 0.91, + "canonical_issue_reason": null, + "canonical_pr_reason": "pull_request:42781 is the more specific and more expansive VibeVoice follow-up, covering the realtime path plus the acoustic tokenizer and additional generation/model work; pull_request:40546 is the earlier base VibeVoice implementation.", + "best_issue_reason": null, + "best_pr_reason": "pull_request:42781 is the better representative of the current VibeVoice work because it covers the broader, newer realtime feature set and includes the shared VibeVoice integration pieces.", + "soft_edge_verdicts": [ + { + "left": "pull_request:40546", + "right": "pull_request:42781", + "accept": false, + "reason": "Reject: both PRs share VibeVoice infrastructure, but 42781 adds a distinct realtime variant and acoustic-tokenizer/model changes, so they are not the same concrete change and would not realistically be merged as one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 3, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-40755-2", + "nodes": [ + "pull_request:40755", + "pull_request:41224" + ], + "soft_pairs": [ + "pull_request:40755|pull_request:41224" + ], + "prepared_review_unit_hash": "69a18bf0dde5addc97f4a6e448a32d2e77172290805e51d6bb21137618e88373", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4675, + "estimated_input_tokens": 1169, + "estimated_eval_tokens": 2594 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4675, + "estimated_input_tokens": 1169, + "estimated_eval_tokens": 2594 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The two PRs are unrelated feature additions in different model families and code paths, so this is not a duplicate cluster.", + "confidence": 0.99, + "canonical_issue_reason": null, + "canonical_pr_reason": "No canonical PR: #40755 adds TimesFM covariate forecasting, while #41224 adds DINOv3 image classification support; they are distinct features in different subsystems.", + "best_issue_reason": null, + "best_pr_reason": "Neither PR is a good global representative for the cluster because they address unrelated enhancements and should not be merged or deduplicated together.", + "soft_edge_verdicts": [ + { + "left": "pull_request:40755", + "right": "pull_request:41224", + "accept": false, + "reason": "Different models and different changes: TimesFM forecasting with covariates vs DINOv3 image classification support. They do not fix the same underlying bug or implement the same change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 4, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-41356-2", + "nodes": [ + "pull_request:41356", + "pull_request:44339" + ], + "soft_pairs": [ + "pull_request:41356|pull_request:44339" + ], + "prepared_review_unit_hash": "aead1308c167b0bae6ff11bf9358340baa5e76f816618383418b505a10a9cd01", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6050, + "estimated_input_tokens": 1513, + "estimated_eval_tokens": 3282 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6050, + "estimated_input_tokens": 1513, + "estimated_eval_tokens": 3282 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Two DEIMv2 pull requests target the same model family, but they are not the same change: PR 41356 is an early/basic model + image processor add, while PR 44339 is a much broader follow-up that adds loss code, auto-mappings, conversion tooling, modular files, and more integration work.", + "confidence": 0.91, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 44339 is the more complete and representative DEIMv2 implementation; PR 41356 looks like an earlier partial version of the same model bring-up.", + "best_issue_reason": null, + "best_pr_reason": "PR 44339 is the strongest single artifact for the cluster because it subsumes the basic model work and includes the broader integration needed for a full Transformers addition.", + "soft_edge_verdicts": [ + { + "left": "pull_request:41356", + "right": "pull_request:44339", + "accept": false, + "reason": "Same model area, but not the same concrete change: 44339 expands the implementation substantially and adds new code paths beyond the basic PR 41356. They look like related iterations, not mergeable duplicates." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 5, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-41797-3", + "nodes": [ + "pull_request:41797", + "pull_request:45186", + "pull_request:45630" + ], + "soft_pairs": [ + "pull_request:41797|pull_request:45186", + "pull_request:45186|pull_request:45630" + ], + "prepared_review_unit_hash": "06859588dc075a9e835878ff034261da2c07225f583000bb155ab791287ef33b", + "component_budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 2, + "serialized_chars": 10409, + "estimated_input_tokens": 2603, + "estimated_eval_tokens": 5462 + }, + "budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 2, + "serialized_chars": 10409, + "estimated_input_tokens": 2603, + "estimated_eval_tokens": 5462 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are three separate new-model integration PRs with overlapping Transformers boilerplate, not the same underlying change. The shared files are generic registration/documentation plumbing, while each PR targets a different model family and code path.", + "confidence": 0.98, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 45186 is the most central representative of the cluster because it connects to both other PRs and reflects the shared model-integration scaffolding.", + "best_issue_reason": null, + "best_pr_reason": "PR 45186 is the best single representative here for triage because it sits at the center of the similarity graph, even though the PRs are not duplicates.", + "soft_edge_verdicts": [ + { + "left": "pull_request:41797", + "right": "pull_request:45186", + "accept": false, + "reason": "Different model additions: DeepSeek OCR vs Isaac. The overlap is only in generic docs/auto-registration files, not the same concrete bug or feature." + }, + { + "left": "pull_request:45186", + "right": "pull_request:45630", + "accept": false, + "reason": "Different model additions: Isaac vs Kimi2-6. Shared filenames are boilerplate integration points, but the PRs implement separate model code paths." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 6, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-42975-2", + "nodes": [ + "pull_request:42975", + "pull_request:42976" + ], + "soft_pairs": [ + "pull_request:42975|pull_request:42976" + ], + "prepared_review_unit_hash": "4e069303b72d97d1faa2436a437e3bea501c536af27852685da957224ed1cc06", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5935, + "estimated_input_tokens": 1484, + "estimated_eval_tokens": 3224 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5935, + "estimated_input_tokens": 1484, + "estimated_eval_tokens": 3224 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These PRs both update GitHub Actions dependencies, but they address different underlying changes: one is a Node 24 compatibility sweep for core workflow actions, while the other is a set of version bumps focused on docker/build-push-action and related workflow updates. They are related in theme, not duplicates.", + "confidence": 0.96, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR #42975 is the better representative of the cluster theme because it targets the core GitHub Actions runtime compatibility problem across many workflows (checkout/cache/upload-artifact).", + "best_issue_reason": null, + "best_pr_reason": "PR #42975 is the stronger representative because it has the broader, more foundational workflow-action compatibility scope; PR #42976 is narrower and mainly upgrades Docker build actions.", + "soft_edge_verdicts": [ + { + "left": "pull_request:42975", + "right": "pull_request:42976", + "accept": false, + "reason": "Do not merge: they upgrade different actions and fix different concerns. #42975 is about Node 24 compatibility for core actions, while #42976 bumps docker/build-push-action and a few workflow versions; they are not the same concrete change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 7, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-42978-7", + "nodes": [ + "pull_request:42978", + "pull_request:43448", + "pull_request:43451", + "pull_request:43973", + "pull_request:44601", + "pull_request:45490", + "pull_request:45838" + ], + "soft_pairs": [ + "pull_request:42978|pull_request:43448", + "pull_request:42978|pull_request:43451", + "pull_request:42978|pull_request:45490", + "pull_request:43448|pull_request:43451", + "pull_request:43448|pull_request:45490", + "pull_request:43451|pull_request:45490", + "pull_request:43973|pull_request:45838", + "pull_request:44601|pull_request:45838", + "pull_request:45490|pull_request:45838" + ], + "prepared_review_unit_hash": "1546ffa79f2dd77710fef19d466c113faab47a9633023d3e0b599dc0b51dcf05", + "component_budget": { + "node_count": 7, + "item_count": 7, + "soft_pair_count": 9, + "serialized_chars": 23542, + "estimated_input_tokens": 5886, + "estimated_eval_tokens": 12028 + }, + "budget": { + "node_count": 7, + "item_count": 7, + "soft_pair_count": 9, + "serialized_chars": 23542, + "estimated_input_tokens": 5886, + "estimated_eval_tokens": 12028 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "This cluster is heterogeneous: it mixes several unrelated model-addition PRs plus one distributed-loading feature PR. Despite superficial filename overlap in docs/auto registries, they do not fix the same underlying code-path problem, so no duplicates should be merged.", + "confidence": 0.98, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "pull_request:42978", + "right": "pull_request:43448", + "accept": false, + "reason": "Different model families and different implementation areas: ViT NEPA vs Molmo. Shared doc/auto registry edits are just boilerplate for new model support, not the same change." + }, + { + "left": "pull_request:42978", + "right": "pull_request:43451", + "accept": false, + "reason": "ViT NEPA and Molmo2 are separate model integrations with different configs, modeling code, and tests; not the same bug or feature." + }, + { + "left": "pull_request:42978", + "right": "pull_request:45490", + "accept": false, + "reason": "ViT NEPA is unrelated to CTSM; both are model additions but for different architectures and code paths." + }, + { + "left": "pull_request:43448", + "right": "pull_request:43451", + "accept": false, + "reason": "Molmo and Molmo2 are distinct model releases with separate packages and tests. Similar naming does not make them the same underlying change." + }, + { + "left": "pull_request:43448", + "right": "pull_request:45490", + "accept": false, + "reason": "Molmo and CTSM are unrelated model additions touching different model code, so they are not duplicates." + }, + { + "left": "pull_request:43451", + "right": "pull_request:45490", + "accept": false, + "reason": "Molmo2 and CTSM are separate model integrations with no shared concrete bug or code-path fix." + }, + { + "left": "pull_request:43973", + "right": "pull_request:45838", + "accept": false, + "reason": "Lfm2 Audio is an audio model addition, while PP-OCRv6 support is text recognition. Different domains, files, and code paths." + }, + { + "left": "pull_request:44601", + "right": "pull_request:45838", + "accept": false, + "reason": "Pipeline-parallel loading support is a distributed systems/runtime change; PP-OCRv6 is a model support PR. They do not address the same problem." + }, + { + "left": "pull_request:45490", + "right": "pull_request:45838", + "accept": false, + "reason": "CTSM model support and PP-OCRv6 text recognition support are distinct model additions with no shared underlying change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 8, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43291-3", + "nodes": [ + "pull_request:43291", + "pull_request:43665", + "pull_request:44815" + ], + "soft_pairs": [ + "pull_request:43291|pull_request:43665", + "pull_request:43291|pull_request:44815", + "pull_request:43665|pull_request:44815" + ], + "prepared_review_unit_hash": "cfc2ce156458e9974119231631ef6442e0b74100a93b7e5dad56c60b6f210e0a", + "component_budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 3, + "serialized_chars": 7093, + "estimated_input_tokens": 1774, + "estimated_eval_tokens": 3804 + }, + "budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 3, + "serialized_chars": 7093, + "estimated_input_tokens": 1774, + "estimated_eval_tokens": 3804 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These PRs are not duplicates: one fixes Whisper decoding/tests, one is a Mistral/FP8 dequantization fix, and one appears to be unrelated test edits in CLIP/ViT. The only plausible representative is the FP8/dequant fix PR.", + "confidence": 0.95, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 44815 is the most substantive and self-contained bug fix, with a clear code-path focus in loading/FP8 quantization and supporting model/test updates.", + "best_issue_reason": null, + "best_pr_reason": "PR 44815 best represents the cluster because it addresses a concrete runtime/path issue across the dequantization and loading flow, unlike the other PRs which are narrower or appear unrelated.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43291", + "right": "pull_request:43665", + "accept": false, + "reason": "Different areas and changes: Whisper tokenizer decoding/tests versus CLIP/ViT test edits. No shared concrete bug or mergeable single fix." + }, + { + "left": "pull_request:43291", + "right": "pull_request:44815", + "accept": false, + "reason": "Whisper decode behavior is unrelated to Mistral FP8/dequant loading logic; they touch different code paths and would not be one combined fix." + }, + { + "left": "pull_request:43665", + "right": "pull_request:44815", + "accept": false, + "reason": "CLIP/ViT test-only edits are unrelated to the FP8 dequantization and model-loading changes in 44815, so these are not the same underlying change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 9, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43488-2", + "nodes": [ + "pull_request:43488", + "pull_request:43532" + ], + "soft_pairs": [ + "pull_request:43488|pull_request:43532" + ], + "prepared_review_unit_hash": "b6fc18938084cf7282db2e5b613fab1104e182c751d08d9cdb8e7118278e154e", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4563, + "estimated_input_tokens": 1141, + "estimated_eval_tokens": 2538 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4563, + "estimated_input_tokens": 1141, + "estimated_eval_tokens": 2538 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Two open PRs are grouped only by weak surface similarity, but they change different code paths: one is a repo-bot/versioning test PR, the other rewires CI/comment-diff workflow plumbing. They are not duplicates.", + "confidence": 0.98, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "pull_request:43488", + "right": "pull_request:43532", + "accept": false, + "reason": "Different underlying changes and goals; shared vit test file is incidental, not evidence of the same fix or mergeable duplicate PRs." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 10, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43612-2", + "nodes": [ + "pull_request:43612", + "pull_request:43613" + ], + "soft_pairs": [ + "pull_request:43612|pull_request:43613" + ], + "prepared_review_unit_hash": "9f8a6b2a2229ff17b1dfa226126422bdbc437821fa6472b56eef90d30bf3b29e", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6083, + "estimated_input_tokens": 1521, + "estimated_eval_tokens": 3298 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6083, + "estimated_input_tokens": 1521, + "estimated_eval_tokens": 3298 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The two pull requests are related feature additions in the same segmentation area, but they implement different pipelines and different code paths: concept segmentation vs visual segmentation. They should not be treated as duplicates.", + "confidence": 0.94, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 43613 is the broader, more integrated feature addition, with more model integrations and test coverage, so it is the better representative PR for this cluster.", + "best_issue_reason": null, + "best_pr_reason": "PR 43613 is the stronger representative because it covers the larger promptable visual segmentation pipeline surface and includes more supporting files and model hookups.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43612", + "right": "pull_request:43613", + "accept": false, + "reason": "Different features and code paths: promptable concept segmentation vs promptable visual segmentation. Shared docs/infrastructure do not make them the same underlying change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 11, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43651-2", + "nodes": [ + "pull_request:43651", + "pull_request:43663" + ], + "soft_pairs": [ + "pull_request:43651|pull_request:43663" + ], + "prepared_review_unit_hash": "f42780f43d885ef13db20945118edd2617314d2499dc3da181cf691e9bbfc30f", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4292, + "estimated_input_tokens": 1073, + "estimated_eval_tokens": 2402 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4292, + "estimated_input_tokens": 1073, + "estimated_eval_tokens": 2402 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are two different trainer.py changes: one addresses gradient-accumulation loss scaling, the other adds an overridable signature-column helper for data filtering. They share a file and subsystem, but not the same underlying bug or mergeable change set.", + "confidence": 0.97, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": "43651 is the more self-contained, user-facing trainer behavior change, but it is not a duplicate of 43663.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43651", + "right": "pull_request:43663", + "accept": false, + "reason": "Different code paths and purposes: loss scaling in training_step vs. column/signature filtering for model inputs. They would not plausibly be merged as one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 12, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43656-3", + "nodes": [ + "pull_request:43656", + "pull_request:43836", + "pull_request:43842" + ], + "soft_pairs": [ + "pull_request:43656|pull_request:43842", + "pull_request:43836|pull_request:43842" + ], + "prepared_review_unit_hash": "8ba016a4f08f2fc6bb99862c3c11efb6636bcedb519e1440e12db066a340f207", + "component_budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 2, + "serialized_chars": 5993, + "estimated_input_tokens": 1499, + "estimated_eval_tokens": 3254 + }, + "budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 2, + "serialized_chars": 5993, + "estimated_input_tokens": 1499, + "estimated_eval_tokens": 3254 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "All three PRs target the same import-time TypeAdapter annotation problem in `src/transformers/cli/serve.py`; PR 43842 is the cleanest, most focused fix.", + "confidence": 0.86, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 43842 is the most focused and explicit fix for the underlying NameError when pydantic is not installed, with a minimal annotation/import change in `serve.py`.", + "best_issue_reason": null, + "best_pr_reason": "PR 43842 best represents the cluster because it directly and narrowly fixes the concrete runtime NameError path without extra unrelated changes.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43656", + "right": "pull_request:43842", + "accept": true, + "reason": "Both PRs address the same import-time NameError from evaluating `TypeAdapter` in `src/transformers/cli/serve.py`; they are compatible fixes for the same code path." + }, + { + "left": "pull_request:43836", + "right": "pull_request:43842", + "accept": true, + "reason": "Both fixes target the same `TypeAdapter` annotation NameError in the CLI serve path; 43836 is broader, but its core change is the same bug fix and could plausibly be merged together." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet: all three PRs touch `src/transformers/cli/serve.py` and are clearly aimed at preventing a `TypeAdapter`/annotation-related import-time NameError. The choice of PR 43842 as the most focused fix is supported by the smaller diff and explicit pydantic gating. The soft-edge accepts are also conservative enough: both pairs appear to be alternative or complementary fixes for the same bug path in the same file, so they look mergeable into a single bugfix PR." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 13, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43779-2", + "nodes": [ + "pull_request:43779", + "pull_request:43816" + ], + "soft_pairs": [ + "pull_request:43779|pull_request:43816" + ], + "prepared_review_unit_hash": "e7c674adcc031aa8883109a5dfc678b4407fb20b10764302092e384d031ba83a", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4362, + "estimated_input_tokens": 1091, + "estimated_eval_tokens": 2438 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4362, + "estimated_input_tokens": 1091, + "estimated_eval_tokens": 2438 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both pull requests appear to implement the same SwanLab integration change: adding run id/resume support in `integration_utils.py` for issue #43698, with one variant also expanding the docs. They are strong duplicate PR candidates.", + "confidence": 0.94, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR #43816 is the cleaner representative: it carries the same SwanLab id/resume support plus the accompanying documentation updates, making it the more complete version of the shared change.", + "best_issue_reason": null, + "best_pr_reason": "PR #43816 is the best global representative because it appears to include the full feature change and docs for SwanLab run id/resume support, while #43779 is the narrower implementation-focused variant.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43779", + "right": "pull_request:43816", + "accept": true, + "reason": "Same underlying SwanLab callback enhancement for id/resume support in the same file and same issue target; the doc differences do not change the core code-path change, so these are duplicate PRs." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet: both PRs target issue #43698, touch the same file, and the previews indicate the same SwanLab id/resume support with #43816 adding docs. The duplicate/mergeable verdict is conservative enough given the shared target and overlapping code change." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 14, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 2, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43996-28", + "nodes": [ + "pull_request:43996", + "pull_request:43997", + "pull_request:44007", + "pull_request:44010", + "pull_request:44013", + "pull_request:44017", + "pull_request:44018", + "pull_request:44019", + "pull_request:44044", + "pull_request:44056", + "pull_request:44059", + "pull_request:44066", + "pull_request:44068", + "pull_request:44071", + "pull_request:44072", + "pull_request:44073", + "pull_request:44074", + "pull_request:44076", + "pull_request:44085", + "pull_request:44086", + "pull_request:44098", + "pull_request:44101", + "pull_request:44116", + "pull_request:44129", + "pull_request:44154", + "pull_request:44161", + "pull_request:44722" + ], + "soft_pairs": [ + "pull_request:43996|pull_request:44085", + "pull_request:43996|pull_request:44044", + "pull_request:44066|pull_request:44085", + "pull_request:44007|pull_request:44072", + "pull_request:44072|pull_request:44722", + "pull_request:44066|pull_request:44072", + "pull_request:44013|pull_request:44044", + "pull_request:44066|pull_request:44086", + "pull_request:44066|pull_request:44071", + "pull_request:44018|pull_request:44068", + "pull_request:44018|pull_request:44066", + "pull_request:44066|pull_request:44068", + "pull_request:44019|pull_request:44722", + "pull_request:44086|pull_request:44722", + "pull_request:44019|pull_request:44085", + "pull_request:44018|pull_request:44019", + "pull_request:44019|pull_request:44071", + "pull_request:44068|pull_request:44116", + "pull_request:44068|pull_request:44722", + "pull_request:44071|pull_request:44722", + "pull_request:44018|pull_request:44116", + "pull_request:44018|pull_request:44086", + "pull_request:44018|pull_request:44085", + "pull_request:44071|pull_request:44085", + "pull_request:44068|pull_request:44086", + "pull_request:44059|pull_request:44086", + "pull_request:44085|pull_request:44086", + "pull_request:44098|pull_request:44101", + "pull_request:44059|pull_request:44722", + "pull_request:44018|pull_request:44071", + "pull_request:44017|pull_request:44722", + "pull_request:44017|pull_request:44161", + "pull_request:44017|pull_request:44073", + "pull_request:44068|pull_request:44071", + "pull_request:44071|pull_request:44086", + "pull_request:43997|pull_request:44073", + "pull_request:44017|pull_request:44071", + "pull_request:44073|pull_request:44101", + "pull_request:44059|pull_request:44071", + "pull_request:44056|pull_request:44101", + "pull_request:44059|pull_request:44068", + "pull_request:44073|pull_request:44161", + "pull_request:44018|pull_request:44059", + "pull_request:44101|pull_request:44161", + "pull_request:43997|pull_request:44017", + "pull_request:43997|pull_request:44056", + "pull_request:44071|pull_request:44076", + "pull_request:44074|pull_request:44101", + "pull_request:44056|pull_request:44161", + "pull_request:43997|pull_request:44018", + "pull_request:43997|pull_request:44161", + "pull_request:44074|pull_request:44161", + "pull_request:44018|pull_request:44161", + "pull_request:43997|pull_request:44071", + "pull_request:44010|pull_request:44071", + "pull_request:44018|pull_request:44129", + "pull_request:44010|pull_request:44073", + "pull_request:43997|pull_request:44074", + "pull_request:43997|pull_request:44129", + "pull_request:44071|pull_request:44129", + "pull_request:44010|pull_request:44018", + "pull_request:43996|pull_request:44161", + "pull_request:43996|pull_request:44017", + "pull_request:44010|pull_request:44044", + "pull_request:44044|pull_request:44074", + "pull_request:44066|pull_request:44722", + "pull_request:44085|pull_request:44722", + "pull_request:44085|pull_request:44129", + "pull_request:43996|pull_request:44007", + "pull_request:43996|pull_request:44129", + "pull_request:44044|pull_request:44129", + "pull_request:44044|pull_request:44154" + ], + "prepared_review_unit_hash": "933fa393f4ee6facf160dcc9f4a64f213051fc64cfba4ee53093a432823501f9", + "component_budget": { + "node_count": 28, + "item_count": 28, + "soft_pair_count": 78, + "serialized_chars": 84895, + "estimated_input_tokens": 21224, + "estimated_eval_tokens": 42704 + }, + "budget": { + "node_count": 27, + "item_count": 27, + "soft_pair_count": 72, + "serialized_chars": 75093, + "estimated_input_tokens": 18774, + "estimated_eval_tokens": 37804 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Mostly independent per-model output-tracing refactors tied to #43979. There are at least two concrete mergeable overlap pockets: GPT-Neo (44018/44068) and GPT-J/CodeGen (44066/44085/44722), but the rest are separate model-specific changes.", + "confidence": 0.9, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 44066 is a reasonable cluster anchor because it is an explicit #43979-linked GPT-J/CodeGen output-tracing refactor and overlaps with the later GPT-J/CodeGen follow-up, but it should be treated as one representative pocket rather than the sole center of the cluster.", + "best_issue_reason": null, + "best_pr_reason": "PR 44066 is the strongest single representative for the cluster because it has an explicit tracking-issue link and covers a broad concrete GPT-J/CodeGen refactor, while the cluster also contains a separate GPT-Neo overlap that should not be collapsed into the same anchor.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43996", + "right": "pull_request:44085", + "accept": false, + "reason": "Same output-tracing initiative, but different model implementations and code paths." + }, + { + "left": "pull_request:43996", + "right": "pull_request:44044", + "accept": false, + "reason": "Both target #43979, but they are different model-specific refactors, not the same change." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44085", + "accept": true, + "reason": "Both touch the same GPT-J output-tracing code path, and these changes could plausibly be merged into one PR." + }, + { + "left": "pull_request:44007", + "right": "pull_request:44072", + "accept": false, + "reason": "Different model families and concrete code paths; only the broad output-tracing theme overlaps." + }, + { + "left": "pull_request:44072", + "right": "pull_request:44722", + "accept": false, + "reason": "Different model implementations; shared tracking-issue context is not enough to make this a duplicate." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44072", + "accept": false, + "reason": "Different models and different output-tracing code paths." + }, + { + "left": "pull_request:44013", + "right": "pull_request:44044", + "accept": false, + "reason": "Different model families and unrelated concrete refactors." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44086", + "accept": false, + "reason": "Both are output-tracing work, but they target different model code paths." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44071", + "accept": false, + "reason": "Different model implementations; same general refactor theme only." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44068", + "accept": true, + "reason": "These are the same GPT-Neo output-tracing refactor in the same model/file family and look mergeable as one PR." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44066", + "accept": false, + "reason": "GPT-Neo vs GPT-J/CodeGen are different concrete code paths." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44068", + "accept": false, + "reason": "GPT-J/CodeGen vs GPT-Neo are separate model-specific changes." + }, + { + "left": "pull_request:44019", + "right": "pull_request:44722", + "accept": false, + "reason": "Different model families and different refactors." + }, + { + "left": "pull_request:44086", + "right": "pull_request:44722", + "accept": false, + "reason": "Different model implementations; the shared theme is too broad to merge them." + }, + { + "left": "pull_request:44019", + "right": "pull_request:44085", + "accept": false, + "reason": "Different model families and output paths." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44019", + "accept": false, + "reason": "GPT-Neo and ResNet are unrelated concrete changes." + }, + { + "left": "pull_request:44019", + "right": "pull_request:44071", + "accept": false, + "reason": "Different models and different output-tracing code paths." + }, + { + "left": "pull_request:44068", + "right": "pull_request:44116", + "accept": false, + "reason": "Different model families and unrelated concrete changes." + }, + { + "left": "pull_request:44068", + "right": "pull_request:44722", + "accept": false, + "reason": "Different model code paths; same initiative only." + }, + { + "left": "pull_request:44071", + "right": "pull_request:44116", + "accept": false, + "reason": "MPT and ImageGPT are different model implementations." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44086", + "accept": false, + "reason": "Different models; shared tracing pattern is too broad to merge." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44085", + "accept": false, + "reason": "GPT-Neo and GPT-J are separate model code paths." + }, + { + "left": "pull_request:44071", + "right": "pull_request:44085", + "accept": false, + "reason": "Different models; only the generic output-tracing theme overlaps." + }, + { + "left": "pull_request:44068", + "right": "pull_request:44086", + "accept": false, + "reason": "Different models and different concrete refactors." + }, + { + "left": "pull_request:44059", + "right": "pull_request:44086", + "accept": false, + "reason": "GPT-2 vs MGP-STR are unrelated concrete changes." + }, + { + "left": "pull_request:44085", + "right": "pull_request:44086", + "accept": false, + "reason": "Different models and code paths." + }, + { + "left": "pull_request:44098", + "right": "pull_request:44101", + "accept": false, + "reason": "Different model families and different fixes." + }, + { + "left": "pull_request:44059", + "right": "pull_request:44722", + "accept": false, + "reason": "Different models; same theme is not enough for duplicate triage." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44071", + "accept": false, + "reason": "GPT-Neo and MPT are different model implementations." + }, + { + "left": "pull_request:44017", + "right": "pull_request:44722", + "accept": false, + "reason": "Different model families and concrete output paths." + }, + { + "left": "pull_request:44017", + "right": "pull_request:44161", + "accept": false, + "reason": "Different model-specific refactors." + }, + { + "left": "pull_request:44017", + "right": "pull_request:44073", + "accept": false, + "reason": "Different model families and code paths." + }, + { + "left": "pull_request:44068", + "right": "pull_request:44071", + "accept": false, + "reason": "Different model implementations; same tracing pattern only." + }, + { + "left": "pull_request:44071", + "right": "pull_request:44086", + "accept": false, + "reason": "Different models and output implementations." + }, + { + "left": "pull_request:43997", + "right": "pull_request:44073", + "accept": false, + "reason": "RegNet and VisualBert are unrelated concrete changes." + }, + { + "left": "pull_request:44017", + "right": "pull_request:44018", + "accept": false, + "reason": "Different model families and output paths." + }, + { + "left": "pull_request:44017", + "right": "pull_request:44056", + "accept": false, + "reason": "Different model families; same tracking initiative only." + }, + { + "left": "pull_request:44071", + "right": "pull_request:44129", + "accept": false, + "reason": "Different models and different output-tracing implementations." + }, + { + "left": "pull_request:44074", + "right": "pull_request:44101", + "accept": false, + "reason": "TextNet and Flaubert are unrelated concrete changes." + }, + { + "left": "pull_request:44056", + "right": "pull_request:44161", + "accept": false, + "reason": "Different models and refactor targets." + }, + { + "left": "pull_request:43997", + "right": "pull_request:44018", + "accept": false, + "reason": "RegNet and GPT-Neo are unrelated concrete code paths." + }, + { + "left": "pull_request:43997", + "right": "pull_request:44161", + "accept": false, + "reason": "Different model families; only the broad output-tracing theme overlaps." + }, + { + "left": "pull_request:44071", + "right": "pull_request:44076", + "accept": false, + "reason": "MPT and ImageGPT are different model-specific changes." + }, + { + "left": "pull_request:44010", + "right": "pull_request:44071", + "accept": false, + "reason": "SqueezeBert and MPT are different model implementations." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44129", + "accept": false, + "reason": "Different models and output code paths." + }, + { + "left": "pull_request:44010", + "right": "pull_request:44073", + "accept": false, + "reason": "Different models; not the same underlying change." + }, + { + "left": "pull_request:44074", + "right": "pull_request:44161", + "accept": false, + "reason": "TextNet vs LongT5 are unrelated concrete changes." + }, + { + "left": "pull_request:43996", + "right": "pull_request:44161", + "accept": false, + "reason": "Different model families and output paths." + }, + { + "left": "pull_request:43996", + "right": "pull_request:44017", + "accept": false, + "reason": "Different model families and concrete code paths." + }, + { + "left": "pull_request:44010", + "right": "pull_request:44044", + "accept": false, + "reason": "Different models; the shared tracing theme is too broad." + }, + { + "left": "pull_request:44044", + "right": "pull_request:44074", + "accept": false, + "reason": "Different models and different fixes." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44722", + "accept": true, + "reason": "Same GPT-J/CodeGen output-tracing family with overlapping GPT-J code; they could plausibly be merged into one PR." + }, + { + "left": "pull_request:44085", + "right": "pull_request:44722", + "accept": true, + "reason": "Both are GPT-J output-tracing changes on the same file/code path, so they look mergeable as one PR." + }, + { + "left": "pull_request:44085", + "right": "pull_request:44129", + "accept": false, + "reason": "GPT-J and SpeechT5 are unrelated model-specific changes." + }, + { + "left": "pull_request:43996", + "right": "pull_request:44007", + "accept": false, + "reason": "Different model families and concrete output paths." + }, + { + "left": "pull_request:43996", + "right": "pull_request:44129", + "accept": false, + "reason": "Different model families; same initiative only." + }, + { + "left": "pull_request:44044", + "right": "pull_request:44129", + "accept": false, + "reason": "DeBERTa and SpeechT5 are unrelated concrete changes." + } + ] + }, + "evaluator_result": { + "accept": false, + "feedback": "The packet supports the general conclusion that this is mostly a set of separate model-by-model output-tracing refactors tied to #43979, but the reasoning overstates the evidence in a few places. In particular, PR 44066 is not the center of the only clearly mergeable overlap: the GPT-Neo pair (44018/44068) is also a strong mergeable overlap, and 44066 also overlaps with 44722. Please keep the verdicts and anchor language more conservative and avoid implying uniqueness where the packet shows multiple mergeable pairs." + }, + "evaluator_used": true, + "retried": true, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 14, + "component_count": 27, + "review_unit_index": 2, + "review_unit_count": 2, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43996-28", + "nodes": [ + "pull_request:43997", + "pull_request:44010", + "pull_request:44024", + "pull_request:44056", + "pull_request:44073", + "pull_request:44074", + "pull_request:44076", + "pull_request:44129" + ], + "soft_pairs": [ + "pull_request:44024|pull_request:44076", + "pull_request:44073|pull_request:44074", + "pull_request:44056|pull_request:44073", + "pull_request:44056|pull_request:44074", + "pull_request:43997|pull_request:44010", + "pull_request:44010|pull_request:44129" + ], + "prepared_review_unit_hash": "edcafb15aa209634d5a49ca54432fc8d284926878e6f33cb4b2ae396bda186ab", + "component_budget": { + "node_count": 28, + "item_count": 28, + "soft_pair_count": 78, + "serialized_chars": 84895, + "estimated_input_tokens": 21224, + "estimated_eval_tokens": 42704 + }, + "budget": { + "node_count": 8, + "item_count": 8, + "soft_pair_count": 6, + "serialized_chars": 17123, + "estimated_input_tokens": 4281, + "estimated_eval_tokens": 8818 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "These PRs are related only by a broad theme: migrating different model implementations to the new standardized output-capture machinery. Each one targets a different model/code path, so they are not duplicates of the same underlying change.", + "confidence": 0.95, + "canonical_issue_reason": null, + "canonical_pr_reason": "No single PR is a safe canonical duplicate target here: every candidate is model-specific and changes a different forward path, even though the refactor pattern is similar.", + "best_issue_reason": null, + "best_pr_reason": "PR 44073 is the clearest representative of the shared refactor pattern, with explicit `capture_outputs` and `can_return_tuple` integration and a straightforward output-tracing migration, but it remains specific to VisualBert rather than a cluster-wide duplicate target.", + "soft_edge_verdicts": [ + { + "left": "pull_request:44024", + "right": "pull_request:44076", + "accept": false, + "reason": "Both are standardized-output refactors, but they touch different models (FocalNet vs ImageGPT) and different forward paths; similar theme, not the same fix." + }, + { + "left": "pull_request:44073", + "right": "pull_request:44074", + "accept": false, + "reason": "Different model implementations (VisualBert vs TextNet) with separate output-capture plumbing; they are not plausibly mergeable into one PR." + }, + { + "left": "pull_request:44056", + "right": "pull_request:44073", + "accept": false, + "reason": "Both adopt output-tracing decorators, but they affect different architectures and code paths (MPNet vs VisualBert), so this is not the same underlying change." + }, + { + "left": "pull_request:44056", + "right": "pull_request:44074", + "accept": false, + "reason": "MPNet and TextNet are unrelated model implementations; the shared refactor pattern is too broad to treat as a duplicate." + }, + { + "left": "pull_request:43997", + "right": "pull_request:44010", + "accept": false, + "reason": "RegNet and SqueezeBert are distinct model stacks with different forward logic; the overlap is only in migration style, not in the same bug or change." + }, + { + "left": "pull_request:44010", + "right": "pull_request:44129", + "accept": false, + "reason": "SqueezeBert and SpeechT5 modify different subsystems and output contracts; they share the decorator migration theme but not a single concrete code-path fix." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 15, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43998-13", + "nodes": [ + "pull_request:43998", + "pull_request:43999", + "pull_request:44000", + "pull_request:44001", + "pull_request:44002", + "pull_request:44003", + "pull_request:44004", + "pull_request:44025", + "pull_request:44026", + "pull_request:44027", + "pull_request:44028", + "pull_request:44029", + "pull_request:44030" + ], + "soft_pairs": [ + "pull_request:43998|pull_request:44002", + "pull_request:43998|pull_request:44000", + "pull_request:43998|pull_request:44001", + "pull_request:43998|pull_request:44003", + "pull_request:43998|pull_request:44029", + "pull_request:43999|pull_request:44001", + "pull_request:43999|pull_request:44002", + "pull_request:43999|pull_request:44025", + "pull_request:43999|pull_request:44026", + "pull_request:43999|pull_request:44027", + "pull_request:44000|pull_request:44001", + "pull_request:44000|pull_request:44002", + "pull_request:44000|pull_request:44003", + "pull_request:44000|pull_request:44025", + "pull_request:44001|pull_request:44002", + "pull_request:44001|pull_request:44003", + "pull_request:44001|pull_request:44025", + "pull_request:44002|pull_request:44029", + "pull_request:44002|pull_request:44003", + "pull_request:44003|pull_request:44028", + "pull_request:44003|pull_request:44004", + "pull_request:43998|pull_request:44004", + "pull_request:44001|pull_request:44004", + "pull_request:44002|pull_request:44004", + "pull_request:44000|pull_request:44004", + "pull_request:44025|pull_request:44026", + "pull_request:44025|pull_request:44027", + "pull_request:44025|pull_request:44028", + "pull_request:44025|pull_request:44029", + "pull_request:44025|pull_request:44030", + "pull_request:44026|pull_request:44029", + "pull_request:44026|pull_request:44027", + "pull_request:44026|pull_request:44028", + "pull_request:44026|pull_request:44030", + "pull_request:44027|pull_request:44028", + "pull_request:44027|pull_request:44029", + "pull_request:44027|pull_request:44030", + "pull_request:44028|pull_request:44029", + "pull_request:44028|pull_request:44030", + "pull_request:44029|pull_request:44030" + ], + "prepared_review_unit_hash": "fac27c8c4c121b855ebaf4ae33eb25c30a640387c3fffe83d14868c548661962", + "component_budget": { + "node_count": 13, + "item_count": 13, + "soft_pair_count": 40, + "serialized_chars": 40167, + "estimated_input_tokens": 10042, + "estimated_eval_tokens": 20340 + }, + "budget": { + "node_count": 13, + "item_count": 13, + "soft_pair_count": 40, + "serialized_chars": 40167, + "estimated_input_tokens": 10042, + "estimated_eval_tokens": 20340 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "This cluster is a set of similar but independent model-specific output-tracing refactors; they share a theme, not a single underlying bug or change.", + "confidence": 0.88, + "canonical_issue_reason": null, + "canonical_pr_reason": "44025 is the best cluster anchor: it\u2019s a straightforward output-tracing refactor with a clear model-specific scope and touches two closely related depth_anything files.", + "best_issue_reason": null, + "best_pr_reason": "44025 is the most representative PR in the set, with a clean refactor scope and the strongest centrality among the similar output-tracing changes.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43998", + "right": "pull_request:44002", + "accept": false, + "reason": "Different models and code paths; same refactor pattern, but not one mergeable change." + }, + { + "left": "pull_request:43998", + "right": "pull_request:44000", + "accept": false, + "reason": "Different model families; this is broad output-tracing cleanup, not the same underlying fix." + }, + { + "left": "pull_request:43998", + "right": "pull_request:44001", + "accept": false, + "reason": "Timm backbone vs UnivNet are unrelated implementations; shared theme is too broad." + }, + { + "left": "pull_request:43998", + "right": "pull_request:44003", + "accept": false, + "reason": "Different subsystems and model stacks; no single concrete bug/change here." + }, + { + "left": "pull_request:43998", + "right": "pull_request:44029", + "accept": false, + "reason": "Different model code paths; similar tracing refactor only." + }, + { + "left": "pull_request:43999", + "right": "pull_request:44001", + "accept": false, + "reason": "MobileNetV1 and UnivNet are unrelated models; not the same PR-worthy change." + }, + { + "left": "pull_request:43999", + "right": "pull_request:44002", + "accept": false, + "reason": "Different model implementations; too broad to treat as duplicates." + }, + { + "left": "pull_request:43999", + "right": "pull_request:44025", + "accept": false, + "reason": "Different model families; same output-tracing theme but separate changes." + }, + { + "left": "pull_request:43999", + "right": "pull_request:44026", + "accept": false, + "reason": "Different code paths and model families; not one concrete fix." + }, + { + "left": "pull_request:43999", + "right": "pull_request:44027", + "accept": false, + "reason": "Model-specific refactors only; these are not mergeable as one PR." + }, + { + "left": "pull_request:44000", + "right": "pull_request:44001", + "accept": false, + "reason": "Vision-text dual encoder vs UnivNet are unrelated; only the refactor pattern matches." + }, + { + "left": "pull_request:44000", + "right": "pull_request:44002", + "accept": false, + "reason": "Different models and behavior; no shared underlying bug." + }, + { + "left": "pull_request:44000", + "right": "pull_request:44003", + "accept": false, + "reason": "Different code paths and architectures; same theme, not same change." + }, + { + "left": "pull_request:44000", + "right": "pull_request:44025", + "accept": false, + "reason": "Separate model implementations; output-tracing refactor is too generic to merge." + }, + { + "left": "pull_request:44001", + "right": "pull_request:44002", + "accept": false, + "reason": "UnivNet and UperNet are unrelated model paths; not the same fix." + }, + { + "left": "pull_request:44001", + "right": "pull_request:44003", + "accept": false, + "reason": "Different architectures and files; similar cleanup only." + }, + { + "left": "pull_request:44001", + "right": "pull_request:44025", + "accept": false, + "reason": "Different models; same refactor style is not enough to duplicate." + }, + { + "left": "pull_request:44002", + "right": "pull_request:44029", + "accept": false, + "reason": "UperNet and RWKV are unrelated; no common concrete bug/change." + }, + { + "left": "pull_request:44002", + "right": "pull_request:44003", + "accept": false, + "reason": "Different model stacks and code paths; not one change." + }, + { + "left": "pull_request:44003", + "right": "pull_request:44028", + "accept": false, + "reason": "Mamba/Falcon-Mamba vs SuperPoint are unrelated implementations; only superficially similar." + }, + { + "left": "pull_request:44003", + "right": "pull_request:44004", + "accept": false, + "reason": "Different model families and forward signatures; not a single mergeable fix." + }, + { + "left": "pull_request:43998", + "right": "pull_request:44004", + "accept": false, + "reason": "Timm backbone and CodeGen are unrelated; same refactor theme is too broad." + }, + { + "left": "pull_request:44001", + "right": "pull_request:44004", + "accept": false, + "reason": "UnivNet vs CodeGen are different code paths; not duplicates." + }, + { + "left": "pull_request:44002", + "right": "pull_request:44004", + "accept": false, + "reason": "Different models and behavior; no shared concrete bug." + }, + { + "left": "pull_request:44000", + "right": "pull_request:44004", + "accept": false, + "reason": "Different model implementations; not the same underlying change." + }, + { + "left": "pull_request:44025", + "right": "pull_request:44026", + "accept": false, + "reason": "Different model wrappers and files; same tracing refactor pattern only." + }, + { + "left": "pull_request:44025", + "right": "pull_request:44027", + "accept": false, + "reason": "DepthAnything vs SpeechEncoderDecoder are unrelated; not one bug fix." + }, + { + "left": "pull_request:44025", + "right": "pull_request:44028", + "accept": false, + "reason": "Different models and outputs; similar refactor, not duplicate." + }, + { + "left": "pull_request:44025", + "right": "pull_request:44029", + "accept": false, + "reason": "DepthAnything and RWKV are unrelated code paths." + }, + { + "left": "pull_request:44025", + "right": "pull_request:44030", + "accept": false, + "reason": "Different model families; same output-tracing cleanup only." + }, + { + "left": "pull_request:44026", + "right": "pull_request:44027", + "accept": false, + "reason": "Encoder-decoder wrappers for different modalities; not the same concrete change." + }, + { + "left": "pull_request:44026", + "right": "pull_request:44028", + "accept": false, + "reason": "VisionEncoderDecoder and SuperPoint are unrelated; only the tracing refactor theme overlaps." + }, + { + "left": "pull_request:44026", + "right": "pull_request:44029", + "accept": false, + "reason": "Different model stacks and behavior; not mergeable as one PR." + }, + { + "left": "pull_request:44026", + "right": "pull_request:44030", + "accept": false, + "reason": "Different model code paths; similar cleanup, not duplicate." + }, + { + "left": "pull_request:44027", + "right": "pull_request:44028", + "accept": false, + "reason": "SpeechEncoderDecoder vs SuperPoint are unrelated implementations." + }, + { + "left": "pull_request:44027", + "right": "pull_request:44029", + "accept": false, + "reason": "Different model families and forward logic; not the same bug/change." + }, + { + "left": "pull_request:44027", + "right": "pull_request:44030", + "accept": false, + "reason": "Separate model-specific refactors; not one concrete PR." + }, + { + "left": "pull_request:44028", + "right": "pull_request:44029", + "accept": false, + "reason": "SuperPoint and RWKV are unrelated; shared refactor pattern is insufficient." + }, + { + "left": "pull_request:44028", + "right": "pull_request:44030", + "accept": false, + "reason": "Different model code paths; not the same underlying change." + }, + { + "left": "pull_request:44029", + "right": "pull_request:44030", + "accept": false, + "reason": "RWKV and DPR are separate models; no single mergeable fix." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 16, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-44662-3", + "nodes": [ + "pull_request:44662", + "pull_request:45075", + "pull_request:45401" + ], + "soft_pairs": [ + "pull_request:44662|pull_request:45401", + "pull_request:44662|pull_request:45075", + "pull_request:45075|pull_request:45401" + ], + "prepared_review_unit_hash": "379b431497882f505caa509e54d6408b81cbd5ce2441305d2c43838eb19b40b6", + "component_budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 3, + "serialized_chars": 10914, + "estimated_input_tokens": 2729, + "estimated_eval_tokens": 5714 + }, + "budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 3, + "serialized_chars": 10914, + "estimated_input_tokens": 2729, + "estimated_eval_tokens": 5714 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These PRs all add separate new model integrations, but for different models and code paths, so they are not duplicates.", + "confidence": 0.98, + "canonical_issue_reason": "No issue items are present in this cluster.", + "canonical_pr_reason": "No canonical PR: PenguinVL, DeepSeek-OCR-2, and Voxtral TTS are distinct model additions, not the same change.", + "best_issue_reason": "No issue candidate to select as best issue.", + "best_pr_reason": "No single best PR for deduplication; each PR implements a different model integration.", + "soft_edge_verdicts": [ + { + "left": "pull_request:44662", + "right": "pull_request:45401", + "accept": false, + "reason": "PenguinVL and Voxtral TTS are different model additions with different architectures and functionality; they do not fix the same underlying problem and would not be merged as one PR." + }, + { + "left": "pull_request:44662", + "right": "pull_request:45075", + "accept": false, + "reason": "PenguinVL and DeepSeek-OCR-2 are separate model implementations; shared registration/docs boilerplate is not enough to treat them as the same change." + }, + { + "left": "pull_request:45075", + "right": "pull_request:45401", + "accept": false, + "reason": "DeepSeek-OCR-2 and Voxtral TTS add different models in different domains, so they are not the same code-path change or duplicate PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 17, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-44979-2", + "nodes": [ + "pull_request:44979", + "pull_request:45363" + ], + "soft_pairs": [ + "pull_request:44979|pull_request:45363" + ], + "prepared_review_unit_hash": "65e585d2f265df15d64ab74be5f85bc21c34ef0ed34ede50226c82a7ad2b027f", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4405, + "estimated_input_tokens": 1102, + "estimated_eval_tokens": 2460 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4405, + "estimated_input_tokens": 1102, + "estimated_eval_tokens": 2460 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These two PRs are related in the broad area of module/kernel fusion, but they are not the same change. PR #44979 adds a generic Module Fusion API and tests, while PR #45363 implements a specific n-to-1 kernel fusion workflow via KernelConfig across multiple integration/modeling files. The overlap is architectural, not a duplicate bug fix or mergeable single patch.", + "confidence": 0.86, + "canonical_issue_reason": null, + "canonical_pr_reason": "#45363 is the more concrete end-user feature implementation: it wires fusion through KernelConfig, touches the actual integration/modeling path, and includes an example. #44979 is the underlying API scaffold, not the same deliverable.", + "best_issue_reason": null, + "best_pr_reason": "#45363 is the best representative PR for the cluster because it contains the practical fusion behavior and integration changes. #44979 is broader infrastructure, but it does not subsume the specific n-to-1 kernel fusion feature.", + "soft_edge_verdicts": [ + { + "left": "pull_request:44979", + "right": "pull_request:45363", + "accept": false, + "reason": "Do not treat as duplicates: #44979 introduces a generic module fusion API, whereas #45363 adds a specific KernelConfig-based n-to-1 kernel fusion implementation. They share a theme but not the same concrete code-path change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 18, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45086-2", + "nodes": [ + "pull_request:45086", + "pull_request:45317" + ], + "soft_pairs": [ + "pull_request:45086|pull_request:45317" + ], + "prepared_review_unit_hash": "24c7c535cdc03a35c40f44017f8591df820e088d770584d2ea6bbb47a11c8ad3", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4414, + "estimated_input_tokens": 1104, + "estimated_eval_tokens": 2464 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4414, + "estimated_input_tokens": 1104, + "estimated_eval_tokens": 2464 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both PRs address the same Mistral regex patch AttributeError in the same code path; 45317 is the fuller version with tests, while 45086 is a smaller earlier variant.", + "confidence": 0.97, + "canonical_issue_reason": null, + "canonical_pr_reason": "45317 is the better canonical PR because it targets the same fix as 45086 but adds test coverage and appears more complete.", + "best_issue_reason": null, + "best_pr_reason": "45317 is the strongest representative of the cluster: same bug fix, same file, same issue target, plus tests verifying the behavior.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45086", + "right": "pull_request:45317", + "accept": true, + "reason": "Same underlying AttributeError fix in _patch_mistral_regex; same file and issue target, and the later PR extends the same change with tests rather than a different fix." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 19, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45115-2", + "nodes": [ + "pull_request:45115", + "pull_request:45569" + ], + "soft_pairs": [ + "pull_request:45115|pull_request:45569" + ], + "prepared_review_unit_hash": "93db06790329c486fa5b959ba9141d15af12809135c950c560a918f329ed9929", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6066, + "estimated_input_tokens": 1517, + "estimated_eval_tokens": 3290 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6066, + "estimated_input_tokens": 1517, + "estimated_eval_tokens": 3290 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both PRs touch NemotronH internals, but they are different changes: one is a small conversion/mapping and inheritance refactor, the other is a broader architectural split into NemotronHDense/NemotronHSparse with docs and tests.", + "confidence": 0.89, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 45569 is the better canonical PR because it represents the larger, more complete NemotronH restructuring, including the new dense/sparse model split, auto-mapping updates, docs, and tests.", + "best_issue_reason": null, + "best_pr_reason": "PR 45569 is the best representative of the cluster since it is the more comprehensive user-facing/model-architecture change; PR 45115 is a narrower preparatory refactor focused on conversion mapping and inheritance wiring.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45115", + "right": "pull_request:45569", + "accept": false, + "reason": "Reject: they share NemotronH-related files, but the concrete changes differ. 45115 is a focused conversion-mapping/refactor PR, while 45569 introduces the dense/sparse model split plus docs/tests. They are related work, not the same underlying change and would not plausibly be merged into one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 20, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45332-2", + "nodes": [ + "pull_request:45332", + "pull_request:45333" + ], + "soft_pairs": [ + "pull_request:45332|pull_request:45333" + ], + "prepared_review_unit_hash": "37732b33b8b5144327a3eccb2d00d78a67834b5a8cf0064fca19fbe7241a90df", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5398, + "estimated_input_tokens": 1350, + "estimated_eval_tokens": 2956 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5398, + "estimated_input_tokens": 1350, + "estimated_eval_tokens": 2956 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Two closely related but distinct PRs: one adds the heterogeneous per-layer config API, and the other builds model/cache support on top of it. They share infrastructure, but they are not the same concrete change.", + "confidence": 0.91, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR #45332 is the better canonical representative because it is the broader umbrella change: it adds heterogeneous model support and also includes the config plumbing it depends on. PR #45333 is a narrower prerequisite subset focused on config serialization/representation.", + "best_issue_reason": null, + "best_pr_reason": "PR #45332 is the strongest single representative of the cluster because it covers the larger end-to-end feature and incorporates the shared heterogeneity foundation used by #45333.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45332", + "right": "pull_request:45333", + "accept": false, + "reason": "Do not treat these as duplicates: #45333 is config-layer support only, while #45332 adds broader model/cache integration and downstream behavior. They are related and sequenced, but not the same underlying code-path fix or mergeable into one PR without collapsing distinct scopes." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 21, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45415-2", + "nodes": [ + "pull_request:45415", + "pull_request:45425" + ], + "soft_pairs": [ + "pull_request:45415|pull_request:45425" + ], + "prepared_review_unit_hash": "764bb3966caa37bafe982d589372f28f8e0b1ee7b003c6edbec50b83d8754ad3", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5514, + "estimated_input_tokens": 1379, + "estimated_eval_tokens": 3014 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5514, + "estimated_input_tokens": 1379, + "estimated_eval_tokens": 3014 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both items are typing-related pull requests, but they make different changes: one is a broad repo-wide type-checking cleanup, the other adds specific typing protocols and related model-utils annotations. They should not be treated as duplicates.", + "confidence": 0.93, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 45425 is the more focused, code-path-specific typing change, adding shared typing protocols and model-utils-related annotations; PR 45415 is a much broader type-checking sweep across many files and a config tweak, so it is not the best representative of the cluster.", + "best_issue_reason": null, + "best_pr_reason": "PR 45425 is the better cluster representative because it introduces concrete typing infrastructure changes in core modules, whereas PR 45415 is mostly a wide annotation/checker cleanup touching many unrelated files.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45415", + "right": "pull_request:45425", + "accept": false, + "reason": "They overlap on typing/theme and share a few filenames, but the underlying changes are different: 45415 is a broad mypy/type-ignore cleanup across many files, while 45425 adds specific typing protocols and model-utils support. Not the same concrete fix/change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 22, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45487-6", + "nodes": [ + "pull_request:45487", + "pull_request:45817", + "pull_request:45843", + "pull_request:45858", + "pull_request:45890", + "pull_request:45926" + ], + "soft_pairs": [ + "pull_request:45487|pull_request:45843", + "pull_request:45487|pull_request:45926", + "pull_request:45817|pull_request:45843", + "pull_request:45858|pull_request:45890", + "pull_request:45858|pull_request:45926", + "pull_request:45890|pull_request:45926" + ], + "prepared_review_unit_hash": "f544d3c5395f37af9059cac4274a03ea3c9a0f0cf8fb4405a565e24bf3797892", + "component_budget": { + "node_count": 6, + "item_count": 6, + "soft_pair_count": 6, + "serialized_chars": 14717, + "estimated_input_tokens": 3680, + "estimated_eval_tokens": 7616 + }, + "budget": { + "node_count": 6, + "item_count": 6, + "soft_pair_count": 6, + "serialized_chars": 14717, + "estimated_input_tokens": 3680, + "estimated_eval_tokens": 7616 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "This cluster mixes several unrelated PRs: two real model-parallel code fixes and three separate XPU test-expectation updates. None of the soft edges should be merged as duplicates.", + "confidence": 0.93, + "canonical_issue_reason": null, + "canonical_pr_reason": "No single PR is a true duplicate of the others, but #45487 is the best representative of the cluster\u2019s substantive bug-fix side because it changes model code to fix device placement in AltCLIP/ChineseCLIP rather than only updating tests.", + "best_issue_reason": null, + "best_pr_reason": "#45487 is the strongest canonical PR candidate overall: it addresses an actual code-path bug in model parallel execution and spans the most directly related affected models, while the others are either different model families or test-only expectation updates.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45487", + "right": "pull_request:45843", + "accept": false, + "reason": "Both mention model parallel, but they fix different models and different code paths (AltCLIP/ChineseCLIP vs Deimv2). Not the same underlying bug." + }, + { + "left": "pull_request:45487", + "right": "pull_request:45926", + "accept": false, + "reason": "#45487 is a source-code device placement fix; #45926 only adds XPU test expectations for HyperCLOVA X. Different models and different change types." + }, + { + "left": "pull_request:45817", + "right": "pull_request:45843", + "accept": false, + "reason": "Both are model-parallel fixes, but for different architectures and distinct implementation issues. They would not plausibly be merged into one PR." + }, + { + "left": "pull_request:45858", + "right": "pull_request:45890", + "accept": false, + "reason": "These are separate test-only XPU expectation updates for different models, not the same bug or change." + }, + { + "left": "pull_request:45858", + "right": "pull_request:45926", + "accept": false, + "reason": "Different model test expectation additions for different architectures; similarity is only at the broad CI/test level." + }, + { + "left": "pull_request:45890", + "right": "pull_request:45926", + "accept": false, + "reason": "Separate model-specific XPU expectation updates; no shared concrete code-path problem." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 23, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45548-2", + "nodes": [ + "pull_request:45548", + "pull_request:45662" + ], + "soft_pairs": [ + "pull_request:45548|pull_request:45662" + ], + "prepared_review_unit_hash": "c192077cd647c9d391cad276140a94397734170337ad7d3c84919bc5856bce08", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4508, + "estimated_input_tokens": 1127, + "estimated_eval_tokens": 2510 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4508, + "estimated_input_tokens": 1127, + "estimated_eval_tokens": 2510 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both PRs are about expert-parallel (EP) loading in Transformers, but they address different failure modes and distributed backends: one is focused on accelerate + DeepSpeed ZeRO-3/config serialization, the other on FSDP2 rank-0 broadcast overwriting expert tensors. They are related but not the same underlying change.", + "confidence": 0.89, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR #45662 is the better canonical representative for the cluster because it targets the core loading path and fixes a concrete EP corruption bug for distributed tensor handling; #45548 is narrower and tied to accelerate/ZeRO-3 config serialization.", + "best_issue_reason": null, + "best_pr_reason": "PR #45662 is the strongest standalone candidate: it fixes a specific, reproducible code-path problem in model loading and has the broader technical scope among the two. #45548 is more specialized and does not cover the same failure mode.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45548", + "right": "pull_request:45662", + "accept": false, + "reason": "Related EP-loading area, but different bugs and backends: #45548 removes distributed_config from serialization and adds has_ep for accelerate/ZeRO-3, while #45662 prevents expert parameters from being overwritten after sharding in FSDP2. They would not naturally merge into one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 24, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45639-2", + "nodes": [ + "pull_request:45639", + "pull_request:45645" + ], + "soft_pairs": [ + "pull_request:45639|pull_request:45645" + ], + "prepared_review_unit_hash": "187a07454dfada8b44baf625e90d9b61819d8c1a2070073e97efa646878254e7", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4589, + "estimated_input_tokens": 1148, + "estimated_eval_tokens": 2552 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4589, + "estimated_input_tokens": 1148, + "estimated_eval_tokens": 2552 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both PRs address the same xdist-related captured_info logging problem; #45645 is the broader superset and likely the better canonical PR.", + "confidence": 0.81, + "canonical_issue_reason": null, + "canonical_pr_reason": "#45645 covers the full end-to-end fix for xdist-safe captured_info artifacts, including the testing utility change plus CI log consumption updates, so it is the more complete canonical PR.", + "best_issue_reason": null, + "best_pr_reason": "#45645 is the best representative because it subsumes the core fix in #45639 and adds the related CI handling needed to preserve and read all captured_info files.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45639", + "right": "pull_request:45645", + "accept": true, + "reason": "Same underlying bug: xdist collisions in captured_info handling for patched testing debug logs. #45645 includes the same testing_utils/test changes and extends them with workflow-side log aggregation, so these look like one coherent fix that could plausibly be merged together." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 25, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45683-2", + "nodes": [ + "pull_request:45683", + "pull_request:45734" + ], + "soft_pairs": [ + "pull_request:45683|pull_request:45734" + ], + "prepared_review_unit_hash": "eba4ca0f26b2d3be0c00e316270a9341d31af2208b44dbab31b1d3a8e81e7538", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4469, + "estimated_input_tokens": 1118, + "estimated_eval_tokens": 2492 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4469, + "estimated_input_tokens": 1118, + "estimated_eval_tokens": 2492 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Two open PRs touch the same quantization helper, but they fix different bugs: one special-cases audio submodules, the other changes how default skip lists are combined with user-supplied skips. They should not be clustered as duplicates.", + "confidence": 0.93, + "canonical_issue_reason": null, + "canonical_pr_reason": "45734 is the better representative PR because it fixes the core skip-module composition logic in `get_modules_to_not_convert`, which is broader and more central than the narrow audio-module special case in 45683.", + "best_issue_reason": null, + "best_pr_reason": "45734 is the strongest standalone PR in the set: it addresses a concrete quantization failure mode in the main code path and its fix is directly about the helper\u2019s behavior rather than a model-specific workaround.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45683", + "right": "pull_request:45734", + "accept": false, + "reason": "Both touch quantization skip logic, but 45683 adds audio-tower/embed_audio exclusions while 45734 changes additive handling of `skip_modules`; these are different underlying bugs and would not naturally merge into one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 26, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45737-2", + "nodes": [ + "pull_request:45737", + "pull_request:45738" + ], + "soft_pairs": [ + "pull_request:45737|pull_request:45738" + ], + "prepared_review_unit_hash": "cc2ade73a49ec6a8883a7a24c433988645588bcd049e8aad2bce56b1b24daa1c", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4131, + "estimated_input_tokens": 1033, + "estimated_eval_tokens": 2322 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4131, + "estimated_input_tokens": 1033, + "estimated_eval_tokens": 2322 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both PRs address the same MusicgenMelody cache initialization bug in `forward()`: replacing `EncoderDecoderCache(DynamicCache, DynamicCache)` with a single `DynamicCache`. PR #45737 is the more complete duplicate because it bundles the fix with a regression test, while #45738 is the fix-only variant.", + "confidence": 0.97, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR #45737 is the best canonical PR because it contains the same code fix as #45738 plus a regression test, making it the most complete representation of the bug fix.", + "best_issue_reason": null, + "best_pr_reason": "PR #45737 is the strongest overall candidate: it fixes the concrete cache bug and adds a targeted regression test, whereas #45738 only changes the implementation.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45737", + "right": "pull_request:45738", + "accept": true, + "reason": "Same underlying code-path bug in `musicgen_melody` cache setup; the diffs match on the core fix and the test-only addition in #45737 could plausibly live in the same PR as #45738." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The assessment is grounded in the packet: both PRs make the same `musicgen_melody` cache initialization change, and #45737 only adds a regression test on top of that fix. The soft-edge accept is conservative enough because the diffs are mergeable into one PR for the same bugfix." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 27, + "component_count": 27, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45851-2", + "nodes": [ + "pull_request:45851", + "pull_request:45947" + ], + "soft_pairs": [ + "pull_request:45851|pull_request:45947" + ], + "prepared_review_unit_hash": "0836afdc4169fa50f00f9ad52c4c9bc6611b7530d4d254c1bb2dcbdcb4bf1e37", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4106, + "estimated_input_tokens": 1027, + "estimated_eval_tokens": 2310 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4106, + "estimated_input_tokens": 1027, + "estimated_eval_tokens": 2310 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are two separate docs-only pull requests in the chat templating area, but they address different topics: template file/storage docs vs. reasoning-field prefill behavior.", + "confidence": 0.93, + "canonical_issue_reason": null, + "canonical_pr_reason": "No canonical PR duplicate here; the patches are about different chat templating documentation changes rather than the same fix.", + "best_issue_reason": null, + "best_pr_reason": "PR 45851 is the better representative because it has an explicit issue target and a broader, more central documentation update.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45851", + "right": "pull_request:45947", + "accept": false, + "reason": "Different changes and different user-facing concerns: one documents chat template storage/writing, the other documents prefill behavior for reasoning fields. They share a subsystem but not the same underlying change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + } + ] +} diff --git a/snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/manifest.json b/snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/manifest.json new file mode 100644 index 0000000000000000000000000000000000000000..360d97c1c2dd31139841e263b48aa6846a1912f4 --- /dev/null +++ b/snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/manifest.json @@ -0,0 +1,14 @@ +{ + "analysis_id": "hybrid-model-20260514t060044z", + "artifacts": { + "hybrid": "snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.json", + "hybrid_reviews": "snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.llm-reviews.json" + }, + "channel": "canonical", + "model": null, + "published_at": "2026-05-14T06:06:24Z", + "repo": "huggingface/transformers", + "schema_version": 1, + "snapshot_id": "20260514T060044Z", + "variant": "hybrid" +} diff --git a/snapshots/20260514T060044Z/manifest.json b/snapshots/20260514T060044Z/manifest.json index 246e4558abdb3b3a298da6c5770aaa8388cfa921..9bbb9e79d34d96b977307afabc7c32592a91c02d 100644 --- a/snapshots/20260514T060044Z/manifest.json +++ b/snapshots/20260514T060044Z/manifest.json @@ -32,6 +32,24 @@ "timeline_events": 405 }, "extracted_at": "2026-05-14T06:00:44Z", + "published_analysis": { + "canonical_analysis_id": "hybrid-model-20260514t060044z", + "runs": { + "hybrid-model-20260514t060044z": { + "analysis_id": "hybrid-model-20260514t060044z", + "artifacts": { + "hybrid": "snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.json", + "hybrid_reviews": "snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/analysis-report-hybrid.llm-reviews.json" + }, + "channel": "canonical", + "manifest_path": "snapshots/20260514T060044Z/analysis-runs/hybrid-model-20260514t060044z/manifest.json", + "model": null, + "published_at": "2026-05-14T06:06:24Z", + "variant": "hybrid" + } + }, + "schema_version": 1 + }, "repo": "huggingface/transformers", "snapshot_id": "20260514T060044Z", "watermark": {