Shalyt commited on
Commit
b23e80c
·
verified ·
1 Parent(s): 2a48cbb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -8,11 +8,12 @@ size_categories:
8
  - 10K<n<100K
9
  task_categories:
10
  - question-answering
 
11
  ---
12
 
13
  # ASyMOB: Algebraic Symbolic Mathematical Operations Benchmark
14
 
15
- Large language models (LLMs) are rapidly approaching the level of proficiency in university-level symbolic mathematics required for applications in advanced science and technology. However, existing benchmarks fall short in assessing the core skills of LLMs in symbolic mathematics—such as integration, limits, differential equations, and algebraic simplification. To address this gap, we introduce ASyMOB, a novel assessment framework focused exclusively on symbolic manipulation, featuring 17,092 unique math challenges, organized by similarity and complexity. ASyMOB enables analysis of LLM failure root-causes and generalization capabilities by comparing performance in problems that differ by simple numerical or symbolic "perturbations". Evaluated LLMs exhibit substantial degradation in performance for all perturbation types (up to -70.3%), suggesting reliance on memorized patterns rather than deeper understanding of symbolic math, even among models achieving high baseline accuracy. Comparing LLM performance to computer algebra systems (CAS, e.g. SymPy), we identify examples where CAS fail while LLMs succeed, as well as problems solved only when combining both approaches. Models capable of integrated code execution yielded higher accuracy compared to their performance without code, particularly stabilizing weaker models (up to +33.1% for certain perturbation types). Notably, the most advanced models (o4-mini, Gemini 2.5 Flash) demonstrate not only high symbolic math proficiency (scoring 96.8% and 97.6% on the unperturbed set), but also remarkable robustness against perturbations, (-21.7% and -21.2% vs. average -50.4% for the other models). This may indicate a recent "phase transition" in the generalization capabilities of frontier LLMs. It remains to be seen whether the path forward lies in deeper integration with more sophisticated external tools, or in developing models so capable that symbolic math systems like CAS become unnecessary.
16
 
17
  # ASyMOB Dataset Generation
18
 
 
8
  - 10K<n<100K
9
  task_categories:
10
  - question-answering
11
+ pretty_name: ASyMOB
12
  ---
13
 
14
  # ASyMOB: Algebraic Symbolic Mathematical Operations Benchmark
15
 
16
+ Large language models (LLMs) are rapidly approaching the level of proficiency in university-level symbolic mathematics required for applications in advanced science and technology. However, existing benchmarks fall short in assessing the core skills of LLMs in symbolic mathematics—such as integration, limits, differential equations, and algebraic simplification. To address this gap, we introduce ASyMOB (pronounced Asimov, in tribute to the renowned author), a novel assessment framework focused exclusively on symbolic manipulation, featuring 17,092 unique math challenges, organized by similarity and complexity. ASyMOB enables analysis of LLM failure root-causes and generalization capabilities by comparing performance in problems that differ by simple numerical or symbolic "perturbations". Evaluated LLMs exhibit substantial degradation in performance for all perturbation types (up to -70.3%), suggesting reliance on memorized patterns rather than deeper understanding of symbolic math, even among models achieving high baseline accuracy. Comparing LLM performance to computer algebra systems (CAS, e.g. SymPy), we identify examples where CAS fail while LLMs succeed, as well as problems solved only when combining both approaches. Models capable of integrated code execution yielded higher accuracy compared to their performance without code, particularly stabilizing weaker models (up to +33.1% for certain perturbation types). Notably, the most advanced models (o4-mini, Gemini 2.5 Flash) demonstrate not only high symbolic math proficiency (scoring 96.8% and 97.6% on the unperturbed set), but also remarkable robustness against perturbations, (-21.7% and -21.2% vs. average -50.4% for the other models). This may indicate a recent "phase transition" in the generalization capabilities of frontier LLMs. It remains to be seen whether the path forward lies in deeper integration with more sophisticated external tools, or in developing models so capable that symbolic math systems like CAS become unnecessary.
17
 
18
  # ASyMOB Dataset Generation
19