maciek-g commited on
Commit
15d0c76
·
verified ·
1 Parent(s): 4b11d0a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -2
README.md CHANGED
@@ -6,19 +6,25 @@ library_name: transformers
6
  ---
7
  # AnalysisObjectTransformer Model
8
 
9
- This repository contains the implementation of the AnalysisObjectTransformer model, a deep learning architecture designed for event classification with reconstructed-object inputs. MultiHeadAttention is used to extract the correlation between jets (hadrons) in the final state. Achieves state-of-the-art performance on final states which can be summarized as jets accompanied by missing transverse energy.
 
 
10
 
11
  ## Model Overview
12
 
13
- The AnalysisObjectTransformer model is structured to process jet-level features (energy, mass, area, btag score) in any order (permutation invariance) and event-level features (angle analysis of missing energy and leading jets) to classify signal from background processes to enhance the sensitivity to rare BSM signatures.
14
 
15
  ### Components
16
 
 
 
 
17
  - **Embedding Layers**: Transform input data into a higher-dimensional space for subsequent processing.
18
  - **Attention Blocks (AttBlock)**: Utilize multi-head attention to capture dependencies between different elements of the input data.
19
  - **Class Blocks (ClassBlock)**: Extend attention mechanisms to incorporate class tokens, enabling the model to focus on class-relevant features. Implementation based on "Going deeper with transformers": https://arxiv.org/abs/2103.17239
20
  - **MLP Head**: A sequence of fully connected layers that maps the output of the transformer blocks to the final prediction targets.
21
 
 
22
  ## Usage
23
 
24
  Firstly, clone the repository:
 
6
  ---
7
  # AnalysisObjectTransformer Model
8
 
9
+ This repository contains the implementation of the AnalysisObjectTransformer model, a deep learning architecture designed for event classification with data from the CERN LHC.
10
+ The model operates reconstructed-object and event-level features. MultiHeadAttention is used to extract the correlation between reconstructed objects such as jets (hadrons) or leptons in the final state, while event-level features capture the event summary, such as total hadronic energy or missing transverse energy.
11
+ Achieves state-of-the-art performance on final states which can be summarized as jets accompanied by missing transverse energy.
12
 
13
  ## Model Overview
14
 
15
+ The AnalysisObjectTransformer model is structured to process object-level features, in the case of jets: energy, mass, area, btag score, in any order (permutation invariance) and event-level features (HT, MET) to classify signal from background processes to enhance the sensitivity to rare BSM signatures.
16
 
17
  ### Components
18
 
19
+
20
+ See [here](https://excalidraw.com/#json=tCXGu1s6Az9wh4md45JU6,A3ezTIoqB10HVxOt4hhRSA) for complete architecure:
21
+
22
  - **Embedding Layers**: Transform input data into a higher-dimensional space for subsequent processing.
23
  - **Attention Blocks (AttBlock)**: Utilize multi-head attention to capture dependencies between different elements of the input data.
24
  - **Class Blocks (ClassBlock)**: Extend attention mechanisms to incorporate class tokens, enabling the model to focus on class-relevant features. Implementation based on "Going deeper with transformers": https://arxiv.org/abs/2103.17239
25
  - **MLP Head**: A sequence of fully connected layers that maps the output of the transformer blocks to the final prediction targets.
26
 
27
+
28
  ## Usage
29
 
30
  Firstly, clone the repository: