Update README.md
Browse files
README.md
CHANGED
|
@@ -9,23 +9,40 @@ pinned: false
|
|
| 9 |
|
| 10 |
# Oxford Reasoning with Machine Learning Lab
|
| 11 |
|
| 12 |
-
|
|
|
|
|
|
|
| 13 |
|
| 14 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
|
| 16 |
## Research Areas
|
| 17 |
|
| 18 |
-
### Benchmarks
|
| 19 |
-
We study the science of LLM evaluation
|
|
|
|
|
|
|
|
|
|
| 20 |
|
| 21 |
-
###
|
| 22 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
|
| 24 |
-
##
|
| 25 |
-
We investigate the risks that advanced AI systems may pose to individuals and society. Our work spans the spectrum of harms, from **bias and toxicity in language models** to **misalignment in agentic systems**, alongside technical methods for mitigation and research on AI governance.
|
| 26 |
|
| 27 |
-
|
| 28 |
-
We conduct large-scale empirical studies of how people use and respond to AI systems in decision-making contexts.
|
| 29 |
|
| 30 |
-
|
| 31 |
-
- Website: [oxrml.com](https://oxrml.com/)
|
|
|
|
| 9 |
|
| 10 |
# Oxford Reasoning with Machine Learning Lab
|
| 11 |
|
| 12 |
+
[](https://oxrml.com)
|
| 13 |
+
[](https://www.ox.ac.uk)
|
| 14 |
+
[](#)
|
| 15 |
|
| 16 |
+
|
| 17 |
+
> *Combining theoretical rigour with empirical investigation to understand how AI models reason, solve complex problems, and collaborate with humans.*
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
---
|
| 21 |
+
|
| 22 |
+
## What We Do
|
| 23 |
+
|
| 24 |
+
We study **how machine learning models process information**, perform tasks, and behave in real-world settings — from the maths of evaluation to the messy reality of human use.
|
| 25 |
+
|
| 26 |
+
---
|
| 27 |
|
| 28 |
## Research Areas
|
| 29 |
|
| 30 |
+
### 📐 Benchmarks & Evaluation
|
| 31 |
+
We study the science of LLM evaluation using systematic reviews, benchmark analysis, and statistical modelling. We develop new benchmarks to test LLM reasoning limits, especially in **adversarial**, **interactive**, and **low-resource language** settings.
|
| 32 |
+
|
| 33 |
+
### 🔬 Agentic AI for Science
|
| 34 |
+
We build agentic systems that automate and augment key stages of the scientific process: literature discovery, evidence synthesis, hypothesis generation, and decision support. Our agents are **reliable**, **transparent**, and grounded in domain expertise.
|
| 35 |
|
| 36 |
+
### 🛡️ AI Safety
|
| 37 |
+
From bias and toxicity to misalignment in agentic systems: we investigate the harms advanced AI may pose to individuals and society, alongside technical mitigation methods and **AI governance** research.
|
| 38 |
+
|
| 39 |
+
### 🤝 Human–AI Interaction
|
| 40 |
+
Large-scale empirical studies of how people use and respond to AI systems in **real-world decision-making** contexts.
|
| 41 |
+
|
| 42 |
+
---
|
| 43 |
|
| 44 |
+
## 🏷️ Topics
|
|
|
|
| 45 |
|
| 46 |
+
`llm-evaluation` `benchmarking` `ai-safety` `agentic-ai` `human-ai-interaction` `reasoning` `nlp` `alignment` `bias` `governance` `low-resource-nlp` `scientific-discovery`
|
|
|
|
| 47 |
|
| 48 |
+
---
|
|
|