Update README.md
Browse files
README.md
CHANGED
|
@@ -6,19 +6,25 @@ library_name: transformers
|
|
| 6 |
---
|
| 7 |
# AnalysisObjectTransformer Model
|
| 8 |
|
| 9 |
-
This repository contains the implementation of the AnalysisObjectTransformer model, a deep learning architecture designed for event classification with
|
|
|
|
|
|
|
| 10 |
|
| 11 |
## Model Overview
|
| 12 |
|
| 13 |
-
The AnalysisObjectTransformer model is structured to process
|
| 14 |
|
| 15 |
### Components
|
| 16 |
|
|
|
|
|
|
|
|
|
|
| 17 |
- **Embedding Layers**: Transform input data into a higher-dimensional space for subsequent processing.
|
| 18 |
- **Attention Blocks (AttBlock)**: Utilize multi-head attention to capture dependencies between different elements of the input data.
|
| 19 |
- **Class Blocks (ClassBlock)**: Extend attention mechanisms to incorporate class tokens, enabling the model to focus on class-relevant features. Implementation based on "Going deeper with transformers": https://arxiv.org/abs/2103.17239
|
| 20 |
- **MLP Head**: A sequence of fully connected layers that maps the output of the transformer blocks to the final prediction targets.
|
| 21 |
|
|
|
|
| 22 |
## Usage
|
| 23 |
|
| 24 |
Firstly, clone the repository:
|
|
|
|
| 6 |
---
|
| 7 |
# AnalysisObjectTransformer Model
|
| 8 |
|
| 9 |
+
This repository contains the implementation of the AnalysisObjectTransformer model, a deep learning architecture designed for event classification with data from the CERN LHC.
|
| 10 |
+
The model operates reconstructed-object and event-level features. MultiHeadAttention is used to extract the correlation between reconstructed objects such as jets (hadrons) or leptons in the final state, while event-level features capture the event summary, such as total hadronic energy or missing transverse energy.
|
| 11 |
+
Achieves state-of-the-art performance on final states which can be summarized as jets accompanied by missing transverse energy.
|
| 12 |
|
| 13 |
## Model Overview
|
| 14 |
|
| 15 |
+
The AnalysisObjectTransformer model is structured to process object-level features, in the case of jets: energy, mass, area, btag score, in any order (permutation invariance) and event-level features (HT, MET) to classify signal from background processes to enhance the sensitivity to rare BSM signatures.
|
| 16 |
|
| 17 |
### Components
|
| 18 |
|
| 19 |
+
|
| 20 |
+
See [here](https://excalidraw.com/#json=tCXGu1s6Az9wh4md45JU6,A3ezTIoqB10HVxOt4hhRSA) for complete architecure:
|
| 21 |
+
|
| 22 |
- **Embedding Layers**: Transform input data into a higher-dimensional space for subsequent processing.
|
| 23 |
- **Attention Blocks (AttBlock)**: Utilize multi-head attention to capture dependencies between different elements of the input data.
|
| 24 |
- **Class Blocks (ClassBlock)**: Extend attention mechanisms to incorporate class tokens, enabling the model to focus on class-relevant features. Implementation based on "Going deeper with transformers": https://arxiv.org/abs/2103.17239
|
| 25 |
- **MLP Head**: A sequence of fully connected layers that maps the output of the transformer blocks to the final prediction targets.
|
| 26 |
|
| 27 |
+
|
| 28 |
## Usage
|
| 29 |
|
| 30 |
Firstly, clone the repository:
|