shreyanshpadarha commited on
Commit
c25e09b
·
verified ·
1 Parent(s): 0a704fe

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +29 -12
README.md CHANGED
@@ -9,23 +9,40 @@ pinned: false
9
 
10
  # Oxford Reasoning with Machine Learning Lab
11
 
12
- We conduct research on **AI evaluation, safety, and human–AI interaction** to advance understanding of how large language models reason, solve complex problems, and collaborate with humans.
 
 
13
 
14
- Our work combines **theoretical rigour** with **empirical investigation** to study how large language models process information, perform tasks, and behave in real-world settings.
 
 
 
 
 
 
 
 
 
 
15
 
16
  ## Research Areas
17
 
18
- ### Benchmarks and Evaluation
19
- We study the science of LLM evaluation, using **systematic reviews**, **benchmark analysis**, and **statistical modelling** to examine the validity of existing evaluation practices. We develop new benchmarks and evaluation frameworks to test the limits of LLM reasoning, especially in **adversarial**, **interactive**, and **low-resource language** settings.
 
 
 
20
 
21
- ### Agentic AI for Science
22
- We build agentic AI systems that automate and augment key stages of the scientific process, including **literature discovery**, **evidence synthesis**, **hypothesis generation**, and **decision support**. A central focus is developing agents that are **reliable**, **transparent**, and **grounded in domain expertise** for real-world scientific and policy applications.
 
 
 
 
 
23
 
24
- ### AI Safety
25
- We investigate the risks that advanced AI systems may pose to individuals and society. Our work spans the spectrum of harms, from **bias and toxicity in language models** to **misalignment in agentic systems**, alongside technical methods for mitigation and research on AI governance.
26
 
27
- ### Human–AI Interaction
28
- We conduct large-scale empirical studies of how people use and respond to AI systems in decision-making contexts.
29
 
30
- ## Links
31
- - Website: [oxrml.com](https://oxrml.com/)
 
9
 
10
  # Oxford Reasoning with Machine Learning Lab
11
 
12
+ [![Website](https://img.shields.io/badge/⚡️_Website-oxrml.com-002147?style=flat-square)](https://oxrml.com)
13
+ [![Oxford](https://img.shields.io/badge/🎓_University-Oxford-002147?style=flat-square)](https://www.ox.ac.uk)
14
+ [![Focus](https://img.shields.io/badge/🔬_Focus-AI_Evaluation_·_Safety_·_Human--AI-c8962e?style=flat-square)](#)
15
 
16
+
17
+ > *Combining theoretical rigour with empirical investigation to understand how AI models reason, solve complex problems, and collaborate with humans.*
18
+
19
+
20
+ ---
21
+
22
+ ## What We Do
23
+
24
+ We study **how machine learning models process information**, perform tasks, and behave in real-world settings — from the maths of evaluation to the messy reality of human use.
25
+
26
+ ---
27
 
28
  ## Research Areas
29
 
30
+ ### 📐 Benchmarks & Evaluation
31
+ We study the science of LLM evaluation using systematic reviews, benchmark analysis, and statistical modelling. We develop new benchmarks to test LLM reasoning limits, especially in **adversarial**, **interactive**, and **low-resource language** settings.
32
+
33
+ ### 🔬 Agentic AI for Science
34
+ We build agentic systems that automate and augment key stages of the scientific process: literature discovery, evidence synthesis, hypothesis generation, and decision support. Our agents are **reliable**, **transparent**, and grounded in domain expertise.
35
 
36
+ ### 🛡️ AI Safety
37
+ From bias and toxicity to misalignment in agentic systems: we investigate the harms advanced AI may pose to individuals and society, alongside technical mitigation methods and **AI governance** research.
38
+
39
+ ### 🤝 Human–AI Interaction
40
+ Large-scale empirical studies of how people use and respond to AI systems in **real-world decision-making** contexts.
41
+
42
+ ---
43
 
44
+ ## 🏷️ Topics
 
45
 
46
+ `llm-evaluation` `benchmarking` `ai-safety` `agentic-ai` `human-ai-interaction` `reasoning` `nlp` `alignment` `bias` `governance` `low-resource-nlp` `scientific-discovery`
 
47
 
48
+ ---