nielsr HF Staff commited on
Commit
a18f76e
·
verified ·
1 Parent(s): 78b95ba

Enhance dataset card: Add paper, code links, task categories, and comprehensive usage

Browse files

This PR significantly improves the `MAC-SLU` dataset card by adding:
- The Hugging Face `paper` link: https://huggingface.co/papers/2512.01603
- The `code` (GitHub) repository link: https://github.com/Gatsby-web/MAC_SLU
- The `task_categories: ['audio-text-to-text']`, `language: ['en']` and `tags: ['spoken-language-understanding', 'automotive', 'multi-intent']` to improve discoverability on the Hub.
- An introductory overview of the dataset.
- A comprehensive `Usage` section, including code snippets for In-Context Learning (ICL) and Supervised Fine-Tuning (SFT), directly from the GitHub README. This helps users quickly get started with the dataset and its associated models.

Files changed (1) hide show
  1. README.md +100 -3
README.md CHANGED
@@ -1,3 +1,100 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ task_categories:
4
+ - audio-text-to-text
5
+ language:
6
+ - en
7
+ tags:
8
+ - spoken-language-understanding
9
+ - automotive
10
+ - multi-intent
11
+ ---
12
+
13
+ # MAC-SLU: A Benchmark for Multi-Intent Spoken Language Understanding in Automotive Cabins
14
+
15
+ [Paper](https://huggingface.co/papers/2512.01603) | [Code](https://github.com/Gatsby-web/MAC_SLU)
16
+
17
+ This repository hosts the **MAC-SLU** dataset, a novel Multi-Intent Automotive Cabin Spoken Language Understanding Benchmark. MAC-SLU is designed to evaluate Spoken Language Understanding (SLU) systems on complex, multi-intent user commands within an automotive environment, addressing the limitations of existing SLU datasets in terms of diversity and complexity. It features authentic and complex multi-intent data, suitable for benchmarking both Large Language Models (LLMs) and Large Audio Language Models (LALMs).
18
+
19
+ ## 🚀 Getting Started
20
+
21
+ ### 1\. Download the Dataset
22
+
23
+ The complete MAC-SLU dataset is hosted on the Hugging Face Hub.
24
+
25
+ * **Dataset Link:** [Gatsby1984/MAC\_SLU](https://huggingface.co/datasets/Gatsby1984/MAC_SLU)
26
+
27
+ ### 2\. Prepare the Environment
28
+
29
+ Our experiments are divided into two main approaches: **In-Context Learning (ICL)** and **Supervised Fine-Tuning (SFT)**. Please set up the appropriate environment for the method you wish to use.
30
+
31
+ ## 🛠️ Usage
32
+
33
+ ### In-Context Learning (ICL)
34
+
35
+ #### Environment Setup
36
+
37
+ Our ICL code relies on `vLLM`. The required version depends on the model you are using. All experiments were conducted with **Python 3.10**.
38
+
39
+ * For **Qwen3** experiments: `pip install vllm==0.9.2`
40
+ * For **Qwen2.5-Omni** experiments: `pip install vllm==0.8.5.post1`
41
+
42
+ #### Running ICL Experiments
43
+
44
+ **Step 1: Deploy the Model with vLLM**
45
+ (This step is not required if you are using a commercial API.)
46
+
47
+ Open a terminal and run the following command to start the vLLM server. This example is for `Qwen2.5-Omni-7B`.
48
+
49
+ ```bash
50
+
51
+ export CUDA_VISIBLE_DEVICES=0
52
+
53
+ vllm serve /path/to/your/Qwen2.5-Omni-7B \
54
+ --served-model-name Qwen2.5-Omni-7B \
55
+ --tensor-parallel-size 1 \
56
+ --gpu-memory-utilization 0.9 \
57
+ --host 0.0.0.0 \
58
+ --port 12355 \
59
+ --uvicorn-log-level warning \
60
+ --disable-log-requests \
61
+ --max-model-len 32768
62
+ ```
63
+
64
+ **Step 2: Run Inference**
65
+
66
+ Once the server is running, open a new terminal and execute the inference script.
67
+
68
+ ```bash
69
+ python slu_icl.py \
70
+ --provider local \
71
+ --input-file /path/to/test_set.jsonl \
72
+ --audio-dir /path/to/audio_test_directory \
73
+ --output-file /path/to/prediction.jsonl \
74
+ --model-name Qwen2.5-Omni-7B \
75
+ --api-base http://0.0.0.0:12355/v1
76
+ ```
77
+
78
+ * **Note:** For other models, you may need to change `--model-name` and the model path in the `vllm serve` command. To use a commercial API, change `--provider` to the appropriate name and configure the necessary API keys.
79
+
80
+ **Step 3: Evaluation**
81
+
82
+ ```bash
83
+ python metrics.py prediction.jsonl icl_label.jsonl
84
+ ```
85
+ -----
86
+
87
+ ### Supervised Fine-Tuning (SFT)
88
+
89
+ #### Environment Setup
90
+
91
+ For SFT experiments, we use the efficient **LLaMA-Factory** framework. Please follow the official instructions to install and set up the environment.
92
+
93
+ * **Framework:** [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory)
94
+
95
+ #### Training Instructions
96
+
97
+ We recommend using a **LoRA-SFT** approach for fine-tuning.
98
+
99
+ 1. **Prepare your dataset** using the format required by LLaMA-Factory.
100
+ 2. **Configure your training run** by selecting a model, dataset, and setting the LoRA hyperparameters.