diff --git "a/2025/DINT Transformer/layout.json" "b/2025/DINT Transformer/layout.json" new file mode 100644--- /dev/null +++ "b/2025/DINT Transformer/layout.json" @@ -0,0 +1,6825 @@ +{ + "pdf_info": [ + { + "para_blocks": [ + { + "bbox": [ + 236, + 73, + 358, + 88 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 236, + 73, + 358, + 88 + ], + "spans": [ + { + "bbox": [ + 236, + 73, + 358, + 88 + ], + "type": "text", + "content": "DINT Transformer" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 107, + 100, + 186, + 114 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 107, + 100, + 186, + 114 + ], + "spans": [ + { + "bbox": [ + 107, + 100, + 186, + 114 + ], + "type": "text", + "content": "Yueyang Cang*" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 262, + 100, + 329, + 114 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 262, + 100, + 329, + 114 + ], + "spans": [ + { + "bbox": [ + 262, + 100, + 329, + 114 + ], + "type": "text", + "content": "Yuhang Liu*" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 404, + 100, + 489, + 115 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 404, + 100, + 489, + 115 + ], + "spans": [ + { + "bbox": [ + 404, + 100, + 489, + 115 + ], + "type": "text", + "content": "Xiaoteng Zhang" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 223, + 115, + 370, + 128 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 223, + 115, + 370, + 128 + ], + "spans": [ + { + "bbox": [ + 223, + 115, + 370, + 128 + ], + "type": "text", + "content": "Tsinghua University / Beijing" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 155, + 129, + 438, + 142 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 155, + 129, + 438, + 142 + ], + "spans": [ + { + "bbox": [ + 155, + 129, + 438, + 142 + ], + "type": "text", + "content": "{cangyy23,yh-liu23,zxt21}@mails.tsinghua.edu.cn" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 169, + 158, + 225, + 169 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 169, + 158, + 225, + 169 + ], + "spans": [ + { + "bbox": [ + 169, + 158, + 225, + 169 + ], + "type": "text", + "content": "Erlu Zhao" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 74, + 172, + 468, + 185 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 74, + 172, + 468, + 185 + ], + "spans": [ + { + "bbox": [ + 74, + 172, + 468, + 185 + ], + "type": "text", + "content": "Peking University Health Science Center / BeiJing Tsinghua University / BeiJing" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 131, + 186, + 264, + 199 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 131, + 186, + 264, + 199 + ], + "spans": [ + { + "bbox": [ + 131, + 186, + 264, + 199 + ], + "type": "text", + "content": "2411210080@bjmu.edu.cn" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 353, + 186, + 440, + 198 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 353, + 186, + 440, + 198 + ], + "spans": [ + { + "bbox": [ + 353, + 186, + 440, + 198 + ], + "type": "text", + "content": "zzushi@126.com" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 155, + 219, + 202, + 232 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 155, + 219, + 202, + 232 + ], + "spans": [ + { + "bbox": [ + 155, + 219, + 202, + 232 + ], + "type": "text", + "content": "Abstract" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 84, + 241, + 274, + 528 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 84, + 241, + 274, + 528 + ], + "spans": [ + { + "bbox": [ + 84, + 241, + 274, + 528 + ], + "type": "text", + "content": "The DIFF Transformer mitigates interference from irrelevant contexts by introducing a differential attention mechanism, thereby enhancing focus on critical tokens. However, this architecture suffers from two major limitations: first, its use of two independent attention matrices leads to numerical instability, and second, it lacks global context modeling, which is essential for identifying globally significant tokens. To address these challenges, we propose the DINT Transformer, which extends the DIFF Transformer by incorporating an integral mechanism. By computing global importance scores and integrating them into the attention matrix, the DINT Transformer not only improves overall numerical stability but also significantly enhances its ability to capture global dependencies. Experimental results demonstrate that the DINT Transformer achieves superior accuracy and robustness across various practical applications, including long-context language modeling and key information retrieval. These advancements establish the DINT Transformer as a highly effective and promising architecture." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 68, + 547, + 154, + 560 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 68, + 547, + 154, + 560 + ], + "spans": [ + { + "bbox": [ + 68, + 547, + 154, + 560 + ], + "type": "text", + "content": "1 Introduction" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 67, + 568, + 291, + 729 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 568, + 291, + 729 + ], + "spans": [ + { + "bbox": [ + 67, + 568, + 291, + 729 + ], + "type": "text", + "content": "Transformer(Vaswani, 2017), as one of the most popular models in the field of artificial intelligence today, is widely used in natural language processing, computer vision, and other fields, especially with the application of decoder-only architectures in large language models (LLMs). Its core lies in the attention mechanism based on softmax, which assigns importance to different tokens in a sequence. However, recent research(Lu et al., 2021) has found that LLMs face the challenge of attention noise when accurately focusing on key information in the context." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 67, + 731, + 291, + 757 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 731, + 291, + 757 + ], + "spans": [ + { + "bbox": [ + 67, + 731, + 291, + 757 + ], + "type": "text", + "content": "To address the issue of attention noise, DIFF Transformer(Ye et al., 2024) introduces a differen" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 302, + 220, + 526, + 341 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 220, + 526, + 341 + ], + "spans": [ + { + "bbox": [ + 302, + 220, + 526, + 341 + ], + "type": "text", + "content": "tial attention mechanism that effectively suppresses the impact of irrelevant context by computing DIFFerence between two independent attention distributions. However, DIFF Transformer still exhibits significant limitations: The use of two independent attention matrices makes it difficult to accurately estimate weights for noisy components, resulting in numerical instability that may adversely affect downstream task performance." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 302, + 342, + 526, + 612 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 342, + 526, + 612 + ], + "spans": [ + { + "bbox": [ + 302, + 342, + 526, + 612 + ], + "type": "text", + "content": "Through our research, we observed that the semantic interpretation of most tokens in a sequence often depends on a few globally critical tokens. Taking sentence processing as an example, key elements such as subjects or main predicate verbs frequently serve as semantic anchors (as illustrated in Figure 1), playing a decisive role in constructing overall meaning. Building on this insight, we developed DINT Transformer by introducing an integral mechanism to extend DIFF Transformer. This integral component computes global importance scores, enabling the model to enhance its focus on critical tokens. Our proposed DINT Transformer not only further reduces attention noise by strengthening the focus on globally important tokens, but also significantly decreases the frequency of negative values in attention matrices through parametric design, thereby improving the model's overall numerical stability and substantially boosting performance." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 302, + 613, + 526, + 774 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 613, + 526, + 774 + ], + "spans": [ + { + "bbox": [ + 302, + 613, + 526, + 774 + ], + "type": "text", + "content": "Through comprehensive experiments on long-context language modeling and key information retrieval tasks, we rigorously validated the efficacy of DINT Transformer. The results demonstrate that DINT Transformer consistently outperforms both conventional Transformer and DIFF Transformer across all tasks. Its integral mechanism not only effectively captures global dependencies and further suppresses attention noise, but also significantly enhances model stability, successfully addressing inherent limitations of existing approaches. Moreover, while maintaining excellent scalability, DINT" + } + ] + } + ], + "index": 17 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 80, + 762, + 156, + 774 + ], + "type": "page_footnote", + "angle": 0, + "lines": [ + { + "bbox": [ + 80, + 762, + 156, + 774 + ], + "spans": [ + { + "bbox": [ + 80, + 762, + 156, + 774 + ], + "type": "text", + "content": "*Equal contribution." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 285, + 780, + 309, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 285, + 780, + 309, + 791 + ], + "spans": [ + { + "bbox": [ + 285, + 780, + 309, + 791 + ], + "type": "text", + "content": "9813" + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 95, + 795, + 497, + 818 + ], + "type": "footer", + "angle": 0, + "lines": [ + { + "bbox": [ + 95, + 795, + 497, + 818 + ], + "spans": [ + { + "bbox": [ + 95, + 795, + 497, + 818 + ], + "type": "text", + "content": "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 9813-9821 November 4-9, 2025 ©2025 Association for Computational Linguistics" + } + ] + } + ], + "index": 20 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 0 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 76, + 104, + 508, + 259 + ], + "blocks": [ + { + "bbox": [ + 79, + 75, + 519, + 103 + ], + "lines": [ + { + "bbox": [ + 79, + 75, + 519, + 103 + ], + "spans": [ + { + "bbox": [ + 79, + 75, + 519, + 103 + ], + "type": "text", + "content": "Einstein was born in Germany and was a physicist, mathematician, and philosopher." + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_caption" + }, + { + "bbox": [ + 76, + 104, + 508, + 259 + ], + "lines": [ + { + "bbox": [ + 76, + 104, + 508, + 259 + ], + "spans": [ + { + "bbox": [ + 76, + 104, + 508, + 259 + ], + "type": "image", + "image_path": "c5832b5874aae07cfe0d6fc9ccf2edcaf9db53c18224229424ae921d100551d7.jpg" + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 66, + 270, + 525, + 320 + ], + "lines": [ + { + "bbox": [ + 66, + 270, + 525, + 320 + ], + "spans": [ + { + "bbox": [ + 66, + 270, + 525, + 320 + ], + "type": "text", + "content": "Figure 1: The DIFF Transformer's use of two independent attention matrices results in a significant proportion of negative values in its final attention scores. In contrast, the DINT Transformer substantially reduces the occurrence of negative values, enhances numerical stability, and more effectively strengthens attention to globally important tokens—such as precisely focusing on key entities like \"Newton\" within sentences." + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "image_caption" + } + ], + "index": 1 + }, + { + "bbox": [ + 67, + 340, + 291, + 422 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 340, + 291, + 422 + ], + "spans": [ + { + "bbox": [ + 67, + 340, + 291, + 422 + ], + "type": "text", + "content": "Transformer delivers substantial performance improvements in downstream tasks such as key information retrieval. These significant findings establish DINT Transformer as a robust foundational architecture for future advancements in sequence modeling and large language models." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 430, + 188, + 443 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 430, + 188, + 443 + ], + "spans": [ + { + "bbox": [ + 67, + 430, + 188, + 443 + ], + "type": "text", + "content": "2 DINT Transformer" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 451, + 291, + 695 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 451, + 291, + 695 + ], + "spans": [ + { + "bbox": [ + 67, + 451, + 291, + 695 + ], + "type": "text", + "content": "DINT Transformer is designed as a robust architecture for sequence modeling, particularly for large language models (LLMs). The model consists of " + }, + { + "bbox": [ + 67, + 451, + 291, + 695 + ], + "type": "inline_equation", + "content": "L" + }, + { + "bbox": [ + 67, + 451, + 291, + 695 + ], + "type": "text", + "content": " stacked layers, where each layer applies a DINT attention module followed by a feedforward network. Starting from token embeddings " + }, + { + "bbox": [ + 67, + 451, + 291, + 695 + ], + "type": "inline_equation", + "content": "X_0 \\in \\mathbb{R}^{N \\times d_{\\mathrm{model}}}" + }, + { + "bbox": [ + 67, + 451, + 291, + 695 + ], + "type": "text", + "content": ", the input is progressively transformed through " + }, + { + "bbox": [ + 67, + 451, + 291, + 695 + ], + "type": "inline_equation", + "content": "L" + }, + { + "bbox": [ + 67, + 451, + 291, + 695 + ], + "type": "text", + "content": " layers to produce the final output " + }, + { + "bbox": [ + 67, + 451, + 291, + 695 + ], + "type": "inline_equation", + "content": "X_L" + }, + { + "bbox": [ + 67, + 451, + 291, + 695 + ], + "type": "text", + "content": ". The key innovation lies in the addition of an integral mechanism within the attention module, which enables effective modeling of global dependencies while preserving numerical stability. The overall structure aligns with common practices, incorporating pre-RMSNorm(Zhang and Sennrich, 2019) and SwiGLU(Ramachandran et al., 2017; Shazeer, 2020) for enhanced performance following LLaMA(Touvron et al., 2023). A diagram of the model architecture is shown in Figure 2." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 67, + 703, + 169, + 715 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 703, + 169, + 715 + ], + "spans": [ + { + "bbox": [ + 67, + 703, + 169, + 715 + ], + "type": "text", + "content": "2.1 DIFF Attention" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 67, + 720, + 291, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 720, + 291, + 775 + ], + "spans": [ + { + "bbox": [ + 67, + 720, + 291, + 775 + ], + "type": "text", + "content": "DIFF attention introduces a differential attention mechanism that reduces attention noise by leveraging the difference between two attention distributions. Specifically, given the input " + }, + { + "bbox": [ + 67, + 720, + 291, + 775 + ], + "type": "inline_equation", + "content": "X \\in \\mathbb{R}^{N \\times d_{\\mathrm{model}}}" + }, + { + "bbox": [ + 67, + 720, + 291, + 775 + ], + "type": "text", + "content": "," + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 302, + 340, + 514, + 354 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 340, + 514, + 354 + ], + "spans": [ + { + "bbox": [ + 302, + 340, + 514, + 354 + ], + "type": "text", + "content": "it is projected to query, key, and value matrices:" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 307, + 362, + 525, + 384 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 307, + 362, + 525, + 384 + ], + "spans": [ + { + "bbox": [ + 307, + 362, + 525, + 384 + ], + "type": "interline_equation", + "content": "[ Q _ {1}; Q _ {2} ] = X W _ {Q}, \\quad [ K _ {1}; K _ {2} ] = X W _ {K}, \\quad V = X W _ {V}, \\tag {1}", + "image_path": "2933593c8b12937e02b47bb68bb9e07c5ce395cc98fa9d80339b937e82d3a59e.jpg" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 384, + 526, + 439 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 384, + 526, + 439 + ], + "spans": [ + { + "bbox": [ + 302, + 384, + 526, + 439 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 302, + 384, + 526, + 439 + ], + "type": "inline_equation", + "content": "Q_{1}, Q_{2}, K_{1}, K_{2} \\in \\mathbb{R}^{N \\times d}" + }, + { + "bbox": [ + 302, + 384, + 526, + 439 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 302, + 384, + 526, + 439 + ], + "type": "inline_equation", + "content": "V \\in \\mathbb{R}^{N \\times 2d}" + }, + { + "bbox": [ + 302, + 384, + 526, + 439 + ], + "type": "text", + "content": " are the projected matrices, and " + }, + { + "bbox": [ + 302, + 384, + 526, + 439 + ], + "type": "inline_equation", + "content": "W_{Q}, W_{K}, W_{V} \\in \\mathbb{R}^{d_{\\mathrm{model}} \\times 2d}" + }, + { + "bbox": [ + 302, + 384, + 526, + 439 + ], + "type": "text", + "content": " are learnable parameters. The differential attention operator computes the output as:" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 313, + 448, + 525, + 480 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 448, + 525, + 480 + ], + "spans": [ + { + "bbox": [ + 313, + 448, + 525, + 480 + ], + "type": "interline_equation", + "content": "\\operatorname {D i f f A t t n} (X) = \\left(\\operatorname {s o f t m a x} \\left(\\frac {Q _ {1} K _ {1} ^ {\\top}}{\\sqrt {d}}\\right) - \\lambda \\cdot \\operatorname {s o f t m a x} \\left(\\frac {Q _ {2} K _ {2} ^ {\\top}}{\\sqrt {d}}\\right)\\right) V \\tag {2}", + "image_path": "4ed96c98fd9c99e50deab56ab83e75c878d70c5a5b3e081b60f0b43f861d95a5.jpg" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 302, + 481, + 525, + 576 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 481, + 525, + 576 + ], + "spans": [ + { + "bbox": [ + 302, + 481, + 525, + 576 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 302, + 481, + 525, + 576 + ], + "type": "inline_equation", + "content": "\\lambda" + }, + { + "bbox": [ + 302, + 481, + 525, + 576 + ], + "type": "text", + "content": " is a learnable scalar parameter. This differential mechanism effectively suppresses irrelevant context, enhancing the robustness of the attention scores by canceling common-mode noise, analogous to the operation of differential amplifiers in electrical engineering. To synchronize learning dynamics, " + }, + { + "bbox": [ + 302, + 481, + 525, + 576 + ], + "type": "inline_equation", + "content": "\\lambda" + }, + { + "bbox": [ + 302, + 481, + 525, + 576 + ], + "type": "text", + "content": " is re-parameterized as:" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 308, + 587, + 525, + 601 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 308, + 587, + 525, + 601 + ], + "spans": [ + { + "bbox": [ + 308, + 587, + 525, + 601 + ], + "type": "interline_equation", + "content": "\\lambda = \\exp \\left(\\lambda_ {q 1} \\cdot \\lambda_ {k 1}\\right) - \\exp \\left(\\lambda_ {q 2} \\cdot \\lambda_ {k 2}\\right) + \\lambda_ {\\text {i n i t}}, \\tag {3}", + "image_path": "8e029276c4f1fd7b4d6a11507e17e145e1856704214651068fb6421fe7d26601.jpg" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 302, + 612, + 525, + 693 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 612, + 525, + 693 + ], + "spans": [ + { + "bbox": [ + 302, + 612, + 525, + 693 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 302, + 612, + 525, + 693 + ], + "type": "inline_equation", + "content": "\\lambda_{q1},\\lambda_{k1},\\lambda_{q2},\\lambda_{k2}\\in \\mathbb{R}^d" + }, + { + "bbox": [ + 302, + 612, + 525, + 693 + ], + "type": "text", + "content": " are learnable vectors, and " + }, + { + "bbox": [ + 302, + 612, + 525, + 693 + ], + "type": "inline_equation", + "content": "\\lambda_{\\mathrm{init}}\\in (0,1)" + }, + { + "bbox": [ + 302, + 612, + 525, + 693 + ], + "type": "text", + "content": " is a constant used for initialization. Empirical results show that setting " + }, + { + "bbox": [ + 302, + 612, + 525, + 693 + ], + "type": "inline_equation", + "content": "\\lambda_{\\mathrm{init}} = 0.8 - 0.6\\times \\exp (-0.3\\cdot (l - 1))" + }, + { + "bbox": [ + 302, + 612, + 525, + 693 + ], + "type": "text", + "content": " , where " + }, + { + "bbox": [ + 302, + 612, + 525, + 693 + ], + "type": "inline_equation", + "content": "l\\in [1,L]" + }, + { + "bbox": [ + 302, + 612, + 525, + 693 + ], + "type": "text", + "content": " represents the layer index, works well in practice." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 302, + 703, + 405, + 714 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 703, + 405, + 714 + ], + "spans": [ + { + "bbox": [ + 302, + 703, + 405, + 714 + ], + "type": "text", + "content": "2.2 DINT Attention" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 302, + 720, + 525, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 720, + 525, + 775 + ], + "spans": [ + { + "bbox": [ + 302, + 720, + 525, + 775 + ], + "type": "text", + "content": "DINT attention extends DIFF attention by introducing an integral mechanism, enhancing the model's ability to capture globally important information while maintaining numerical stability through row" + } + ] + } + ], + "index": 16 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "spans": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "type": "text", + "content": "9814" + } + ] + } + ], + "index": 17 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 1 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 95, + 72, + 303, + 297 + ], + "blocks": [ + { + "bbox": [ + 95, + 72, + 303, + 297 + ], + "lines": [ + { + "bbox": [ + 95, + 72, + 303, + 297 + ], + "spans": [ + { + "bbox": [ + 95, + 72, + 303, + 297 + ], + "type": "image", + "image_path": "31ad5755d7c15747522595c809d40b7a5a348b94dbbb6239259f152dd0b25d9d.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 67, + 305, + 525, + 329 + ], + "lines": [ + { + "bbox": [ + 67, + 305, + 525, + 329 + ], + "spans": [ + { + "bbox": [ + 67, + 305, + 525, + 329 + ], + "type": "text", + "content": "Figure 2: Multi-head DINT Attention. DIFF Attention matrix implements reducing attention noise, while the Integration Attention matrix enhances global attention." + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "type": "image", + "bbox": [ + 303, + 68, + 503, + 295 + ], + "blocks": [ + { + "bbox": [ + 303, + 68, + 503, + 295 + ], + "lines": [ + { + "bbox": [ + 303, + 68, + 503, + 295 + ], + "spans": [ + { + "bbox": [ + 303, + 68, + 503, + 295 + ], + "type": "image", + "image_path": "7fe408db740e492ca4c65c75575a10fb090b5c9e260f8c7dede65700cf7ff21c.jpg" + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_body" + } + ], + "index": 1 + }, + { + "type": "image", + "bbox": [ + 68, + 339, + 277, + 504 + ], + "blocks": [ + { + "bbox": [ + 68, + 339, + 277, + 504 + ], + "lines": [ + { + "bbox": [ + 68, + 339, + 277, + 504 + ], + "spans": [ + { + "bbox": [ + 68, + 339, + 277, + 504 + ], + "type": "image", + "image_path": "480685ce65be66fd92181b7ad59764bbfb94fd2ac8eaca3c5519cf0f37bc2c52.jpg" + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 69, + 511, + 274, + 523 + ], + "lines": [ + { + "bbox": [ + 69, + 511, + 274, + 523 + ], + "spans": [ + { + "bbox": [ + 69, + 511, + 274, + 523 + ], + "type": "text", + "content": "(a) Scaling model size ranging from 830M to 13B." + } + ] + } + ], + "index": 4, + "angle": 0, + "type": "image_caption" + } + ], + "index": 3 + }, + { + "type": "image", + "bbox": [ + 317, + 339, + 526, + 504 + ], + "blocks": [ + { + "bbox": [ + 317, + 339, + 526, + 504 + ], + "lines": [ + { + "bbox": [ + 317, + 339, + 526, + 504 + ], + "spans": [ + { + "bbox": [ + 317, + 339, + 526, + 504 + ], + "type": "image", + "image_path": "9a08374fb58333f784fbb45c631712cfff30c532a9f971665336402eb27bc4c6.jpg" + } + ] + } + ], + "index": 5, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 317, + 511, + 524, + 523 + ], + "lines": [ + { + "bbox": [ + 317, + 511, + 524, + 523 + ], + "spans": [ + { + "bbox": [ + 317, + 511, + 524, + 523 + ], + "type": "text", + "content": "(b) Scaling number of training tokens for 3B models." + } + ] + } + ], + "index": 6, + "angle": 0, + "type": "image_caption" + }, + { + "bbox": [ + 67, + 533, + 525, + 581 + ], + "lines": [ + { + "bbox": [ + 67, + 533, + 525, + 581 + ], + "spans": [ + { + "bbox": [ + 67, + 533, + 525, + 581 + ], + "type": "text", + "content": "Figure 3: Language modeling loss of scaling up parameter count and training tokens. DINT Transformer outperforms other models, demonstrating that it requires fewer parameters or tokens to achieve comparable performance. (a) DINT Transformer matches the performance of larger models with fewer parameters. (b) DINT Transformer achieves comparable performance using significantly fewer training tokens." + } + ] + } + ], + "index": 7, + "angle": 0, + "type": "image_caption" + } + ], + "index": 5 + }, + { + "bbox": [ + 67, + 602, + 290, + 643 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 602, + 290, + 643 + ], + "spans": [ + { + "bbox": [ + 67, + 602, + 290, + 643 + ], + "type": "text", + "content": "normalization in the final attention matrix. The signal attention matrix " + }, + { + "bbox": [ + 67, + 602, + 290, + 643 + ], + "type": "inline_equation", + "content": "A_{1}" + }, + { + "bbox": [ + 67, + 602, + 290, + 643 + ], + "type": "text", + "content": " is computed using " + }, + { + "bbox": [ + 67, + 602, + 290, + 643 + ], + "type": "inline_equation", + "content": "Q_{1}" + }, + { + "bbox": [ + 67, + 602, + 290, + 643 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 67, + 602, + 290, + 643 + ], + "type": "inline_equation", + "content": "K_{1}" + }, + { + "bbox": [ + 67, + 602, + 290, + 643 + ], + "type": "text", + "content": ":" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 118, + 658, + 290, + 689 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 118, + 658, + 290, + 689 + ], + "spans": [ + { + "bbox": [ + 118, + 658, + 290, + 689 + ], + "type": "interline_equation", + "content": "A _ {1} = \\operatorname {s o f t m a x} \\left(\\frac {Q _ {1} K _ {1} ^ {\\top}}{\\sqrt {d}}\\right). \\tag {4}", + "image_path": "bcc9c7712d63fb814a51ba51f142e40ef93f5085ad247be35b751cc0893e2316.jpg" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 67, + 708, + 291, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 708, + 291, + 775 + ], + "spans": [ + { + "bbox": [ + 67, + 708, + 291, + 775 + ], + "type": "text", + "content": "The integral component computes global importance scores by column-wise averaging of the signal attention weights. Crucially, to prevent information leakage, the averaging operation only considers tokens preceding the current token in the" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 303, + 603, + 349, + 615 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 303, + 603, + 349, + 615 + ], + "spans": [ + { + "bbox": [ + 303, + 603, + 349, + 615 + ], + "type": "text", + "content": "sequence." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 355, + 620, + 525, + 655 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 355, + 620, + 525, + 655 + ], + "spans": [ + { + "bbox": [ + 355, + 620, + 525, + 655 + ], + "type": "interline_equation", + "content": "G [ n,: ] = \\frac {1}{n} \\sum_ {m = 1} ^ {n} A _ {1} [ m,: ], \\tag {5}", + "image_path": "4b5599a3c21bce6f76f83dfaeed72055d7d0768ee4206e61b4228aa5db40c852.jpg" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 302, + 660, + 389, + 672 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 660, + 389, + 672 + ], + "spans": [ + { + "bbox": [ + 302, + 660, + 389, + 672 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 302, + 660, + 389, + 672 + ], + "type": "inline_equation", + "content": "G\\in \\mathbb{R}^{N\\times N}" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 312, + 674, + 525, + 687 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 312, + 674, + 525, + 687 + ], + "spans": [ + { + "bbox": [ + 312, + 674, + 525, + 687 + ], + "type": "text", + "content": "DINT attention operator computes the output as:" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 309, + 692, + 525, + 719 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 309, + 692, + 525, + 719 + ], + "spans": [ + { + "bbox": [ + 309, + 692, + 525, + 719 + ], + "type": "interline_equation", + "content": "\\operatorname {D I N T A t t n} (X) = \\left(A _ {\\text {d i f f}} + \\gamma \\cdot \\operatorname {s o f t m a x} (G)\\right) V, \\tag {6}", + "image_path": "cc86bf87c093cd2af42f64e794bf8a12a39ac52378fd81d02e8a01590d31774d.jpg" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 302, + 721, + 525, + 747 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 721, + 525, + 747 + ], + "spans": [ + { + "bbox": [ + 302, + 721, + 525, + 747 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 302, + 721, + 525, + 747 + ], + "type": "inline_equation", + "content": "\\gamma" + }, + { + "bbox": [ + 302, + 721, + 525, + 747 + ], + "type": "text", + "content": " is a learnable scalar following DIFF Transformer, " + }, + { + "bbox": [ + 302, + 721, + 525, + 747 + ], + "type": "inline_equation", + "content": "A_{\\mathrm{diff}}" + }, + { + "bbox": [ + 302, + 721, + 525, + 747 + ], + "type": "text", + "content": " is DIFF attention component." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 302, + 748, + 524, + 773 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 748, + 524, + 773 + ], + "spans": [ + { + "bbox": [ + 302, + 748, + 524, + 773 + ], + "type": "text", + "content": "Unified Parameter Setting. By setting " + }, + { + "bbox": [ + 302, + 748, + 524, + 773 + ], + "type": "inline_equation", + "content": "\\lambda" + }, + { + "bbox": [ + 302, + 748, + 524, + 773 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 302, + 748, + 524, + 773 + ], + "type": "inline_equation", + "content": "\\gamma" + }, + { + "bbox": [ + 302, + 748, + 524, + 773 + ], + "type": "text", + "content": " to the same value, we ensure that the final attention" + } + ] + } + ], + "index": 17 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "spans": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "type": "text", + "content": "9815" + } + ] + } + ], + "index": 18 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 2 + }, + { + "para_blocks": [ + { + "bbox": [ + 67, + 71, + 290, + 153 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 71, + 290, + 153 + ], + "spans": [ + { + "bbox": [ + 67, + 71, + 290, + 153 + ], + "type": "text", + "content": "matrix " + }, + { + "bbox": [ + 67, + 71, + 290, + 153 + ], + "type": "inline_equation", + "content": "A_{\\mathrm{final}}" + }, + { + "bbox": [ + 67, + 71, + 290, + 153 + ], + "type": "text", + "content": " has rows that sum to 1. This row normalization guarantees numerical stability and consistency across the model, thusintaining data stability throughout the layers. This unified setting follows the parameterization method used in DIFF Transformer, further enhancing training stability." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 161, + 254, + 173 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 161, + 254, + 173 + ], + "spans": [ + { + "bbox": [ + 67, + 161, + 254, + 173 + ], + "type": "text", + "content": "2.3 Multi-Head Differential Attention" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 67, + 178, + 289, + 274 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 178, + 289, + 274 + ], + "spans": [ + { + "bbox": [ + 67, + 178, + 289, + 274 + ], + "type": "text", + "content": "We also use the multi-head mechanism in DINT Transformer. Let " + }, + { + "bbox": [ + 67, + 178, + 289, + 274 + ], + "type": "inline_equation", + "content": "h" + }, + { + "bbox": [ + 67, + 178, + 289, + 274 + ], + "type": "text", + "content": " denote the number of attention heads. We use different projection matrices " + }, + { + "bbox": [ + 67, + 178, + 289, + 274 + ], + "type": "inline_equation", + "content": "W_{Q}^{i}, W_{K}^{i}, W_{V}^{i}, i \\in [1, h]" + }, + { + "bbox": [ + 67, + 178, + 289, + 274 + ], + "type": "text", + "content": " for the heads. The scalar " + }, + { + "bbox": [ + 67, + 178, + 289, + 274 + ], + "type": "inline_equation", + "content": "\\lambda" + }, + { + "bbox": [ + 67, + 178, + 289, + 274 + ], + "type": "text", + "content": " is shared between heads within the same layer. Then the head outputs are normalized and projected to the final results as follows:" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 83, + 285, + 290, + 301 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 83, + 285, + 290, + 301 + ], + "spans": [ + { + "bbox": [ + 83, + 285, + 290, + 301 + ], + "type": "interline_equation", + "content": "\\operatorname {h e a d} _ {i} = \\operatorname {D i f f A t t n} \\left(X; W _ {Q} ^ {i}, W _ {K} ^ {i}, W _ {V} ^ {i}, \\lambda\\right) \\tag {7}", + "image_path": "77b29398709fb3beb60ef0fbb099d5e0b4aaef513c9063782a7d031b3b58c28a.jpg" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 133, + 311, + 289, + 325 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 133, + 311, + 289, + 325 + ], + "spans": [ + { + "bbox": [ + 133, + 311, + 289, + 325 + ], + "type": "interline_equation", + "content": "\\overline {{\\operatorname {h e a d} _ {i}}} = \\operatorname {L N} \\left(\\operatorname {h e a d} _ {i}\\right) \\tag {8}", + "image_path": "1b2956ddbaf9ce423c7c6efe53931e7698885fe82862c2f9762123c4b03133c3.jpg" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 331, + 289, + 358 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 331, + 289, + 358 + ], + "spans": [ + { + "bbox": [ + 69, + 331, + 289, + 358 + ], + "type": "inline_equation", + "content": "\\operatorname{MultiHead}(X) = \\operatorname{Concat}(\\overline{\\operatorname{head}}_1, \\dots, \\overline{\\operatorname{head}}_h) W_O" + }, + { + "bbox": [ + 69, + 331, + 289, + 358 + ], + "type": "text", + "content": " (9)" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "spans": [ + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "inline_equation", + "content": "W_O \\in \\mathbb{R}^{d_{\\mathrm{model}} \\times d_{\\mathrm{model}}}" + }, + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "text", + "content": " is a learnable projection matrix, " + }, + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "inline_equation", + "content": "\\mathrm{LN}(\\cdot)" + }, + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "text", + "content": " uses RMSNorm for each head, and " + }, + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "inline_equation", + "content": "\\mathrm{Concat}(\\cdot)" + }, + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "text", + "content": " concatenates the heads together along the channel dimension. Unlike DIFF Transformer, we do not apply an additional multiplier to the outputs of each head, as the unified parameter setting in DINT Transformer already ensures numerical stability and consistency. The number of heads is set as " + }, + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "inline_equation", + "content": "h = d_{\\mathrm{model}} / 2d" + }, + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "text", + "content": ", where " + }, + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "inline_equation", + "content": "d" + }, + { + "bbox": [ + 67, + 359, + 290, + 520 + ], + "type": "text", + "content": " is the head dimension of the Transformer, to ensure that the parameter count and computational complexity are aligned." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 67, + 521, + 291, + 670 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 521, + 291, + 670 + ], + "spans": [ + { + "bbox": [ + 67, + 521, + 291, + 670 + ], + "type": "text", + "content": "Headwise Normalization. Figure 2 illustrates the use of GroupNorm(Wu and He, 2018) within the attention mechanism to stabilize training. Although Layer Normalization (LN) is applied independently to each attention head, the sparse nature of differential attention often leads to varied statistical patterns across heads. By normalizing each head individually before the concatenation step, LN ensures more consistent gradient statistics, which contributes to improved training stability(Qin et al., 2022; Wang et al., 2023)." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 67, + 679, + 193, + 691 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 679, + 193, + 691 + ], + "spans": [ + { + "bbox": [ + 67, + 679, + 193, + 691 + ], + "type": "text", + "content": "2.4 Overall Architecture" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 67, + 696, + 289, + 750 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 696, + 289, + 750 + ], + "spans": [ + { + "bbox": [ + 67, + 696, + 289, + 750 + ], + "type": "text", + "content": "The overall architecture stacks " + }, + { + "bbox": [ + 67, + 696, + 289, + 750 + ], + "type": "inline_equation", + "content": "L" + }, + { + "bbox": [ + 67, + 696, + 289, + 750 + ], + "type": "text", + "content": " layers, where each layer contains a multihead differential attention module and a feedforward network module. We describe DINT Transformer layer as:" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 104, + 760, + 289, + 775 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 104, + 760, + 289, + 775 + ], + "spans": [ + { + "bbox": [ + 104, + 760, + 289, + 775 + ], + "type": "interline_equation", + "content": "Y ^ {l} = \\operatorname {M u l t i H e a d} \\left(\\mathrm {L N} \\left(X ^ {l}\\right)\\right) + X ^ {l} \\tag {10}", + "image_path": "8a02c6dba5736cbce954647002605973900130b2462d06c9ba737ff5b5e18d6c.jpg" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 339, + 69, + 525, + 84 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 339, + 69, + 525, + 84 + ], + "spans": [ + { + "bbox": [ + 339, + 69, + 525, + 84 + ], + "type": "interline_equation", + "content": "X ^ {l + 1} = \\operatorname {S w i G L U} \\left(\\operatorname {L N} \\left(Y ^ {l}\\right)\\right) + Y ^ {l} \\tag {11}", + "image_path": "01c62f646967e7c3004e319975577c5701fb775e4adefc3574f4e3a12d56430e.jpg" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 302, + 91, + 524, + 117 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 91, + 524, + 117 + ], + "spans": [ + { + "bbox": [ + 302, + 91, + 524, + 117 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 302, + 91, + 524, + 117 + ], + "type": "inline_equation", + "content": "\\mathrm{LN}(\\cdot)" + }, + { + "bbox": [ + 302, + 91, + 524, + 117 + ], + "type": "text", + "content": " is RMSNorm, and " + }, + { + "bbox": [ + 302, + 91, + 524, + 117 + ], + "type": "inline_equation", + "content": "\\operatorname{SwiGLU}(X)" + }, + { + "bbox": [ + 302, + 91, + 524, + 117 + ], + "type": "text", + "content": " is defined as:" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 312, + 129, + 514, + 143 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 312, + 129, + 514, + 143 + ], + "spans": [ + { + "bbox": [ + 312, + 129, + 514, + 143 + ], + "type": "interline_equation", + "content": "\\operatorname {S w i G L U} (X) = \\left(\\operatorname {s w i s h} \\left(X W _ {G}\\right) \\odot X W _ {1}\\right) W _ {2},", + "image_path": "2176604921e8a5f58be2c07e1573bf7c0492134a0d94a81dbb4232fc0b8e1ad6.jpg" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 302, + 154, + 524, + 182 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 154, + 524, + 182 + ], + "spans": [ + { + "bbox": [ + 302, + 154, + 524, + 182 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 302, + 154, + 524, + 182 + ], + "type": "inline_equation", + "content": "W_{G}, W_{1} \\in \\mathbb{R}^{d_{\\mathrm{model}} \\times \\frac{8}{3} d_{\\mathrm{model}}}" + }, + { + "bbox": [ + 302, + 154, + 524, + 182 + ], + "type": "text", + "content": ", and " + }, + { + "bbox": [ + 302, + 154, + 524, + 182 + ], + "type": "inline_equation", + "content": "W_{2} \\in \\mathbb{R}^{\\frac{8}{3} d_{\\mathrm{model}} \\times d_{\\mathrm{model}}}" + }, + { + "bbox": [ + 302, + 154, + 524, + 182 + ], + "type": "text", + "content": " are learnable matrices." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 302, + 194, + 390, + 206 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 194, + 390, + 206 + ], + "spans": [ + { + "bbox": [ + 302, + 194, + 390, + 206 + ], + "type": "text", + "content": "3 Experiments" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 301, + 215, + 526, + 432 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 301, + 215, + 526, + 432 + ], + "spans": [ + { + "bbox": [ + 301, + 215, + 526, + 432 + ], + "type": "text", + "content": "In this study, we evaluate DINT Transformer through a series of experiments, comparing it with DIFF Transformer and other baseline models. Since DINT Transformer does not introduce new learnable parameters, only increasing computational complexity, its parameter count remains unchanged. Therefore, the model configurations used in the comparison were chosen to be the same as those of DIFF Transformer. Our experiments show that by enhancing attention to globally significant tokens, DINT Transformer effectively reduces attention noise. Additionally, DINT Transformer exhibits stronger stability compared to DIFF Transformer, leading to improved performance across tasks such as long-sequence modeling, key information retrieval, and in-context learning." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 302, + 441, + 478, + 454 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 441, + 478, + 454 + ], + "spans": [ + { + "bbox": [ + 302, + 441, + 478, + 454 + ], + "type": "text", + "content": "3.1 Language Modeling Evaluation" + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 302, + 459, + 524, + 512 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 459, + 524, + 512 + ], + "spans": [ + { + "bbox": [ + 302, + 459, + 524, + 512 + ], + "type": "text", + "content": "We trained a 3B DINT Transformer language model using the same configuration settings as the 3B DIFF Transformer language model. The model settings are shown in Table 1." + } + ] + } + ], + "index": 18 + }, + { + "type": "table", + "bbox": [ + 345, + 521, + 483, + 688 + ], + "blocks": [ + { + "bbox": [ + 345, + 521, + 483, + 688 + ], + "lines": [ + { + "bbox": [ + 345, + 521, + 483, + 688 + ], + "spans": [ + { + "bbox": [ + 345, + 521, + 483, + 688 + ], + "type": "table", + "html": "
ParamsValues
Layers28
Hidden size3072
FFN size8192
Vocab size100,288
Heads12
Adam β(0.9, 0.95)
LR3.2 × 10-4
Batch size4M
Warmup steps1000
Weight decay0.1
Dropout0.0
", + "image_path": "ba2c1257118fe419a1df8a2b81b7e6a25b1912b8d2a3f364cec1e2626a734cd4.jpg" + } + ] + } + ], + "index": 19, + "angle": 0, + "type": "table_body" + } + ], + "index": 19 + }, + { + "bbox": [ + 302, + 696, + 524, + 719 + ], + "lines": [ + { + "bbox": [ + 302, + 696, + 524, + 719 + ], + "spans": [ + { + "bbox": [ + 302, + 696, + 524, + 719 + ], + "type": "text", + "content": "Table 1: Configuration settings used for the 3B-size DINT Transformer and DIFF Transformer models." + } + ] + } + ], + "index": 20, + "angle": 0, + "type": "text" + }, + { + "bbox": [ + 302, + 735, + 525, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 735, + 525, + 775 + ], + "spans": [ + { + "bbox": [ + 302, + 735, + 525, + 775 + ], + "type": "text", + "content": "Results. Table 2 presents the zero-shot evaluation results on the LM Eval Harness benchmark (Gao et al., 2023). We compare DINT Transformer" + } + ] + } + ], + "index": 21 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "spans": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "type": "text", + "content": "9816" + } + ] + } + ], + "index": 22 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 3 + }, + { + "para_blocks": [ + { + "bbox": [ + 67, + 71, + 291, + 194 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 71, + 291, + 194 + ], + "spans": [ + { + "bbox": [ + 67, + 71, + 291, + 194 + ], + "type": "text", + "content": "with other state-of-the-art Transformer-based models, including OpenLLaMA-v2-3B (Geng and Liu, 2023), StableLM-base-alpha-3B-v2 (Tow, 2023), and StableLM-3B-4E1T (Tow et al., 2023). All models were trained on 1T tokens under identical configurations to ensure fair comparison. The results demonstrate that DINT Transformer not only outperforms these baselines across all downstream tasks but also exhibits superior stability." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 202, + 283, + 216 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 202, + 283, + 216 + ], + "spans": [ + { + "bbox": [ + 67, + 202, + 283, + 216 + ], + "type": "text", + "content": "3.2 Scalability Compared with Transformer" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 67, + 220, + 290, + 327 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 220, + 290, + 327 + ], + "spans": [ + { + "bbox": [ + 67, + 220, + 290, + 327 + ], + "type": "text", + "content": "We evaluated the scalability of DINT Transformer compared to the standard Transformer and DIFF Transformer, specifically focusing on language modeling tasks. This evaluation involved scaling both model size and the number of training tokens. We adopted an enhanced Transformer architecture similar to LLaMA, ensuring a fair comparison by using identical experimental setups." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 67, + 328, + 291, + 476 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 328, + 291, + 476 + ], + "spans": [ + { + "bbox": [ + 67, + 328, + 291, + 476 + ], + "type": "text", + "content": "Scaling Model Size As shown in Figure 3(a), DINT Transformer consistently outperformed both Transformer and DIFF Transformer across various model sizes (see Table 3 for model configurations). Specifically, DINT Transformer achieved comparable validation loss to the Transformer with " + }, + { + "bbox": [ + 67, + 328, + 291, + 476 + ], + "type": "inline_equation", + "content": "44\\%" + }, + { + "bbox": [ + 67, + 328, + 291, + 476 + ], + "type": "text", + "content": " fewer parameters and matched the performance of DIFF Transformer with " + }, + { + "bbox": [ + 67, + 328, + 291, + 476 + ], + "type": "inline_equation", + "content": "29\\%" + }, + { + "bbox": [ + 67, + 328, + 291, + 476 + ], + "type": "text", + "content": " fewer parameters. This demonstrates the superior efficiency and scalability of DINT Transformer in terms of parameter usage." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 67, + 478, + 291, + 613 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 478, + 291, + 613 + ], + "spans": [ + { + "bbox": [ + 67, + 478, + 291, + 613 + ], + "type": "text", + "content": "Scaling Training Tokens Figure 3(b) shows the results of scaling the number of training tokens. The fitted curves indicate that DINT Transformer achieved comparable performance to the Transformer with " + }, + { + "bbox": [ + 67, + 478, + 291, + 613 + ], + "type": "inline_equation", + "content": "33\\%" + }, + { + "bbox": [ + 67, + 478, + 291, + 613 + ], + "type": "text", + "content": " fewer training tokens. Additionally, DINT Transformer outperformed DIFF Transformer with " + }, + { + "bbox": [ + 67, + 478, + 291, + 613 + ], + "type": "inline_equation", + "content": "16\\%" + }, + { + "bbox": [ + 67, + 478, + 291, + 613 + ], + "type": "text", + "content": " fewer training tokens. These results highlight the significant data efficiency of DINT Transformer, achieving equivalent or superior results with considerably fewer resources." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 622, + 220, + 634 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 622, + 220, + 634 + ], + "spans": [ + { + "bbox": [ + 67, + 622, + 220, + 634 + ], + "type": "text", + "content": "3.3 Key Information Retrieval" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 67, + 640, + 290, + 734 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 640, + 290, + 734 + ], + "spans": [ + { + "bbox": [ + 67, + 640, + 290, + 734 + ], + "type": "text", + "content": "The Needle-In-A-Haystack test (Kamradt, 2023) is used to evaluate the ability of models to extract key information from long contexts. Following the protocol of LWM (Liu et al., 2024) and Gemini 1.5 (Reid et al., 2024), \"needles\" are short sentences that assign a unique number to a city. The objective is to retrieve these numbers based on a given query." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 67, + 735, + 290, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 735, + 290, + 775 + ], + "spans": [ + { + "bbox": [ + 67, + 735, + 290, + 775 + ], + "type": "text", + "content": "We position the answer needle at different depths within the context (0%, 25%, 50%, 75%, 100%), while other needles are placed randomly. Each" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 302, + 71, + 526, + 111 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 71, + 526, + 111 + ], + "spans": [ + { + "bbox": [ + 302, + 71, + 526, + 111 + ], + "type": "text", + "content": "combination of depth and context length is evaluated over 50 samples, and the average accuracy is reported." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 302, + 111, + 526, + 260 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 111, + 526, + 260 + ], + "spans": [ + { + "bbox": [ + 302, + 111, + 526, + 260 + ], + "type": "text", + "content": "Retrieve from 4K Context Length We evaluated the multi-needle retrieval task using 4K-length contexts, inserting " + }, + { + "bbox": [ + 302, + 111, + 526, + 260 + ], + "type": "inline_equation", + "content": "N = 1,2,4,6" + }, + { + "bbox": [ + 302, + 111, + 526, + 260 + ], + "type": "text", + "content": " needles and retrieving " + }, + { + "bbox": [ + 302, + 111, + 526, + 260 + ], + "type": "inline_equation", + "content": "R = 1,2" + }, + { + "bbox": [ + 302, + 111, + 526, + 260 + ], + "type": "text", + "content": " needles. The models used for evaluation were trained with an input length of 4K. As shown in Table 4, DINT Transformer consistently outperforms the other models. Particularly at " + }, + { + "bbox": [ + 302, + 111, + 526, + 260 + ], + "type": "inline_equation", + "content": "N = 6,R = 2" + }, + { + "bbox": [ + 302, + 111, + 526, + 260 + ], + "type": "text", + "content": ", DINT achieves an accuracy of 0.88, significantly better than Transformer and DIFF models, indicating its superior ability to retrieve key information amidst distracting contexts." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "spans": [ + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "text", + "content": "Retrieve from 64K Context Length As shown in Figure 4, the context lengths evaluated range from 8K to 64K, with the configuration set to " + }, + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "inline_equation", + "content": "N = 8" + }, + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "inline_equation", + "content": "R = 1" + }, + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "text", + "content": ". We evaluated the 3B-scale model with extended context (as described in Section 3.3). The accuracy is reported across different answer needle depths (y-axis) and context lengths (x-axis). The bottom row shows the average accuracy across all depths. From the figure, it can be observed that DINT Transformer consistently performs well across varying context lengths and needle depths. Notably, at a 40K context length and " + }, + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "inline_equation", + "content": "25\\%" + }, + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "text", + "content": " needle depth, DINT Transformer shows a " + }, + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "inline_equation", + "content": "52\\%" + }, + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "text", + "content": " improvement in accuracy compared to Transformer and a " + }, + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "inline_equation", + "content": "12\\%" + }, + { + "bbox": [ + 302, + 261, + 526, + 464 + ], + "type": "text", + "content": " improvement compared to DIFF Transformer." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 302, + 465, + 525, + 612 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 465, + 525, + 612 + ], + "spans": [ + { + "bbox": [ + 302, + 465, + 525, + 612 + ], + "type": "text", + "content": "Attention Score Analysis Table 5 presents the attention scores assigned to the answer span and the noise context in the key information retrieval task. These scores reflect the model's ability to focus on relevant information while ignoring irrelevant noise. We compare the normalized attention scores for different depths (i.e., positions) of the target answer within the context. The results show that DINT Transformer allocates significantly higher attention to the correct answer span and exhibits a substantial reduction in attention noise." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 302, + 622, + 428, + 635 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 622, + 428, + 635 + ], + "spans": [ + { + "bbox": [ + 302, + 622, + 428, + 635 + ], + "type": "text", + "content": "3.4 In-Context Learning" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 302, + 640, + 525, + 734 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 640, + 525, + 734 + ], + "spans": [ + { + "bbox": [ + 302, + 640, + 525, + 734 + ], + "type": "text", + "content": "We investigate in-context learning from two main angles: the performance on many-shot classification tasks and the model's ability to maintain robustness when utilizing context. In-context learning is an essential trait of language models, reflecting their capability to make effective use of the provided input context." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 302, + 735, + 525, + 774 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 735, + 525, + 774 + ], + "spans": [ + { + "bbox": [ + 302, + 735, + 525, + 774 + ], + "type": "text", + "content": "Many-Shot In-Context Learning As presented in Figure 5, we compare the accuracy of many-shot classification between DIFF Transformer and" + } + ] + } + ], + "index": 14 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "spans": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "type": "text", + "content": "9817" + } + ] + } + ], + "index": 15 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 4 + }, + { + "para_blocks": [ + { + "type": "table", + "bbox": [ + 69, + 68, + 527, + 138 + ], + "blocks": [ + { + "bbox": [ + 69, + 68, + 527, + 138 + ], + "lines": [ + { + "bbox": [ + 69, + 68, + 527, + 138 + ], + "spans": [ + { + "bbox": [ + 69, + 68, + 527, + 138 + ], + "type": "table", + "html": "
ModelARC-CARC-EBoolQHellaSwagOBQAPIQAWinoGrandeAvg
OpenLLaMA-3B-v233.967.665.770.026.676.762.957.5
StableLM-base-alpha-3B-v232.467.364.668.627.176.063.057.0
StableLM-3B-4E1T-66.6---76.863.2-
DIFF-3B36.9 ± 2.172.6 ± 1.769.2 ± 1.871.1 ± 2.429.1 ± 0.876.5 ± 1.069.2 ± 2.060.6
DINT-3B39.2 ± 1.774.3 ± 1.370.7 ± 1.272.6 ± 1.730.3 ± 0.577.3 ± 0.672.0 ± 1.262.3
", + "image_path": "10a7a295bc1982ed77033c0e7973ac91fc35fd7a574e82bd81017cc34a75adc2.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "table_body" + } + ], + "index": 0 + }, + { + "bbox": [ + 67, + 144, + 525, + 170 + ], + "lines": [ + { + "bbox": [ + 67, + 144, + 525, + 170 + ], + "spans": [ + { + "bbox": [ + 67, + 144, + 525, + 170 + ], + "type": "text", + "content": "Table 2: Eval Harness accuracy compared with well-trained Transformer language models. The results indicate the superior performance of DINT Transformer over other models across a range of tasks." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "text" + }, + { + "type": "image", + "bbox": [ + 89, + 178, + 505, + 312 + ], + "blocks": [ + { + "bbox": [ + 89, + 178, + 505, + 312 + ], + "lines": [ + { + "bbox": [ + 89, + 178, + 505, + 312 + ], + "spans": [ + { + "bbox": [ + 89, + 178, + 505, + 312 + ], + "type": "image", + "image_path": "65797ce23fe06f57dab39f6766fae2cb20d0451f232d71c8b7501ec43c1ff022.jpg" + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 187, + 318, + 404, + 332 + ], + "lines": [ + { + "bbox": [ + 187, + 318, + 404, + 332 + ], + "spans": [ + { + "bbox": [ + 187, + 318, + 404, + 332 + ], + "type": "text", + "content": "Figure 4: Multi-needle retrieval results in 64K length." + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "image_caption" + } + ], + "index": 2 + }, + { + "type": "image", + "bbox": [ + 104, + 364, + 291, + 512 + ], + "blocks": [ + { + "bbox": [ + 104, + 364, + 291, + 512 + ], + "lines": [ + { + "bbox": [ + 104, + 364, + 291, + 512 + ], + "spans": [ + { + "bbox": [ + 104, + 364, + 291, + 512 + ], + "type": "image", + "image_path": "34def1c886986bfdf2ed6c21043ec26f7630e459c2700ec8815cf001773a4b4b.jpg" + } + ] + } + ], + "index": 4, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 145, + 515, + 249, + 526 + ], + "lines": [ + { + "bbox": [ + 145, + 515, + 249, + 526 + ], + "spans": [ + { + "bbox": [ + 145, + 515, + 249, + 526 + ], + "type": "text", + "content": "(a) TREC dataset (6 classes)" + } + ] + } + ], + "index": 5, + "angle": 0, + "type": "image_caption" + } + ], + "index": 4 + }, + { + "type": "image", + "bbox": [ + 301, + 364, + 489, + 513 + ], + "blocks": [ + { + "bbox": [ + 301, + 364, + 489, + 513 + ], + "lines": [ + { + "bbox": [ + 301, + 364, + 489, + 513 + ], + "spans": [ + { + "bbox": [ + 301, + 364, + 489, + 513 + ], + "type": "image", + "image_path": "2a3a072d61bca3ef79a89f07338c944c07f1d8b8da998d7b7d494fee68b701b3.jpg" + } + ] + } + ], + "index": 6, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 333, + 515, + 458, + 526 + ], + "lines": [ + { + "bbox": [ + 333, + 515, + 458, + 526 + ], + "spans": [ + { + "bbox": [ + 333, + 515, + 458, + 526 + ], + "type": "text", + "content": "(b) TREC-fine dataset (50 classes)" + } + ] + } + ], + "index": 7, + "angle": 0, + "type": "image_caption" + } + ], + "index": 6 + }, + { + "type": "image", + "bbox": [ + 105, + 538, + 291, + 687 + ], + "blocks": [ + { + "bbox": [ + 105, + 538, + 291, + 687 + ], + "lines": [ + { + "bbox": [ + 105, + 538, + 291, + 687 + ], + "spans": [ + { + "bbox": [ + 105, + 538, + 291, + 687 + ], + "type": "image", + "image_path": "dfaaa09edd7b40873909753120f4e32211571bd4b7e4be95d12b01c270c9131d.jpg" + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 133, + 692, + 261, + 703 + ], + "lines": [ + { + "bbox": [ + 133, + 692, + 261, + 703 + ], + "spans": [ + { + "bbox": [ + 133, + 692, + 261, + 703 + ], + "type": "text", + "content": "(c) Banking-77 dataset (77 classes)" + } + ] + } + ], + "index": 9, + "angle": 0, + "type": "image_caption" + } + ], + "index": 8 + }, + { + "type": "image", + "bbox": [ + 302, + 539, + 489, + 688 + ], + "blocks": [ + { + "bbox": [ + 302, + 539, + 489, + 688 + ], + "lines": [ + { + "bbox": [ + 302, + 539, + 489, + 688 + ], + "spans": [ + { + "bbox": [ + 302, + 539, + 489, + 688 + ], + "type": "image", + "image_path": "740297a760cc81126bbf6e43d425e3cfbdeab7b1782fb8cb0228f921c6e8eaeb.jpg" + } + ] + } + ], + "index": 10, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 331, + 692, + 460, + 703 + ], + "lines": [ + { + "bbox": [ + 331, + 692, + 460, + 703 + ], + "spans": [ + { + "bbox": [ + 331, + 692, + 460, + 703 + ], + "type": "text", + "content": "(d) Clinic-150 dataset (150 classes)" + } + ] + } + ], + "index": 11, + "angle": 0, + "type": "image_caption" + }, + { + "bbox": [ + 67, + 713, + 525, + 748 + ], + "lines": [ + { + "bbox": [ + 67, + 713, + 525, + 748 + ], + "spans": [ + { + "bbox": [ + 67, + 713, + 525, + 748 + ], + "type": "text", + "content": "Figure 5: Accuracy of many-shot in-context learning across four datasets, with demonstration examples increasing from 1-shot up to a total of 64K tokens. The dashed lines indicate the average accuracy once the model's performance stabilizes." + } + ] + } + ], + "index": 12, + "angle": 0, + "type": "image_caption" + } + ], + "index": 10 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 286, + 781, + 309, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 286, + 781, + 309, + 791 + ], + "spans": [ + { + "bbox": [ + 286, + 781, + 309, + 791 + ], + "type": "text", + "content": "9818" + } + ] + } + ], + "index": 13 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 5 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 66, + 68, + 291, + 233 + ], + "blocks": [ + { + "bbox": [ + 66, + 68, + 291, + 233 + ], + "lines": [ + { + "bbox": [ + 66, + 68, + 291, + 233 + ], + "spans": [ + { + "bbox": [ + 66, + 68, + 291, + 233 + ], + "type": "image", + "image_path": "a1ecdebe228fec7fb1e1ba60fa89203a44d762688b66f9150a21f07f26b4b063.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 103, + 239, + 254, + 252 + ], + "lines": [ + { + "bbox": [ + 103, + 239, + 254, + 252 + ], + "spans": [ + { + "bbox": [ + 103, + 239, + 254, + 252 + ], + "type": "text", + "content": "(a) Examples are randomly arranged." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "type": "image", + "bbox": [ + 302, + 67, + 526, + 232 + ], + "blocks": [ + { + "bbox": [ + 302, + 67, + 526, + 232 + ], + "lines": [ + { + "bbox": [ + 302, + 67, + 526, + 232 + ], + "spans": [ + { + "bbox": [ + 302, + 67, + 526, + 232 + ], + "type": "image", + "image_path": "e10a43e12a60e35af6dff3d0a00fdde35b52e9e2d3008842f0f2ef78b1777c75.jpg" + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 320, + 239, + 509, + 252 + ], + "lines": [ + { + "bbox": [ + 320, + 239, + 509, + 252 + ], + "spans": [ + { + "bbox": [ + 320, + 239, + 509, + 252 + ], + "type": "text", + "content": "(b) Examples are arranged alternately by class." + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "image_caption" + }, + { + "bbox": [ + 67, + 261, + 525, + 298 + ], + "lines": [ + { + "bbox": [ + 67, + 261, + 525, + 298 + ], + "spans": [ + { + "bbox": [ + 67, + 261, + 525, + 298 + ], + "type": "text", + "content": "Figure 6: Many-shot in-context learning accuracy on four datasets. The accuracy for both DIFF Transformer and DINT (Ours) models is presented, showing performance improvements across different numbers of demonstration samples." + } + ] + } + ], + "index": 4, + "angle": 0, + "type": "image_caption" + } + ], + "index": 2 + }, + { + "type": "table", + "bbox": [ + 74, + 317, + 283, + 400 + ], + "blocks": [ + { + "bbox": [ + 74, + 317, + 283, + 400 + ], + "lines": [ + { + "bbox": [ + 74, + 317, + 283, + 400 + ], + "spans": [ + { + "bbox": [ + 74, + 317, + 283, + 400 + ], + "type": "table", + "html": "
SizeHidden Dim.#Layers#Heads
830M1536248
1.4B2048248
2.8B25603210
6.8B40963216
13.1B51204020
", + "image_path": "074b62a6a21bd38d397acb1194075522f23751838a2c90ae6eddfcbd54c3b60b.jpg" + } + ] + } + ], + "index": 5, + "angle": 0, + "type": "table_body" + } + ], + "index": 5 + }, + { + "type": "table", + "bbox": [ + 76, + 483, + 283, + 545 + ], + "blocks": [ + { + "bbox": [ + 67, + 409, + 291, + 470 + ], + "lines": [ + { + "bbox": [ + 67, + 409, + 291, + 470 + ], + "spans": [ + { + "bbox": [ + 67, + 409, + 291, + 470 + ], + "type": "text", + "content": "Table 3: Model configurations for different sizes, including hidden dimension, number of layers, and number of attention heads. Each model was trained with a sequence length of 2048 and a batch size of 0.25 million tokens, for a total of 40K training steps." + } + ] + } + ], + "index": 6, + "angle": 0, + "type": "table_caption" + }, + { + "bbox": [ + 76, + 483, + 283, + 545 + ], + "lines": [ + { + "bbox": [ + 76, + 483, + 283, + 545 + ], + "spans": [ + { + "bbox": [ + 76, + 483, + 283, + 545 + ], + "type": "table", + "html": "
ModelN=1N=2N=4N=6
R=1R=2R=2R=2
Transformer1.000.850.620.55
DIFF1.000.920.840.85
DINT1.000.960.890.88
", + "image_path": "0450518cf89ee4fdef0c68b5a04d2e5e5a496e25ff86f387edc801d5950968b3.jpg" + } + ] + } + ], + "index": 7, + "angle": 0, + "type": "table_body" + } + ], + "index": 7 + }, + { + "bbox": [ + 67, + 554, + 290, + 602 + ], + "lines": [ + { + "bbox": [ + 67, + 554, + 290, + 602 + ], + "spans": [ + { + "bbox": [ + 67, + 554, + 290, + 602 + ], + "type": "text", + "content": "Table 4: Multi-needle retrieval accuracy in 4K length contexts, averaged over the answer needle positions. " + }, + { + "bbox": [ + 67, + 554, + 290, + 602 + ], + "type": "inline_equation", + "content": "N" + }, + { + "bbox": [ + 67, + 554, + 290, + 602 + ], + "type": "text", + "content": " represents the number of needles, and " + }, + { + "bbox": [ + 67, + 554, + 290, + 602 + ], + "type": "inline_equation", + "content": "R" + }, + { + "bbox": [ + 67, + 554, + 290, + 602 + ], + "type": "text", + "content": " denotes the number of query cities." + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "text" + }, + { + "bbox": [ + 67, + 626, + 292, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 626, + 292, + 775 + ], + "spans": [ + { + "bbox": [ + 67, + 626, + 292, + 775 + ], + "type": "text", + "content": "our DINT Transformer architecture. We evaluate the 3B-size language models that support 64K input length. We follow the evaluation protocol of (Bertsch et al., 2024) and use constrained decoding (Ratner et al., 2023). The number of demonstration samples is incrementally increased from 1-shot until the total length reaches 64K tokens. Specifically, we evaluate the models on the following datasets: TREC (Hovy et al., 2001) with 50 classes, Banking-77 (Casanueva et al., 2020) with 77 classes, and Clinic-150 (Larson et al., 2019) with 150 classes." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 319, + 526, + 413 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 319, + 526, + 413 + ], + "spans": [ + { + "bbox": [ + 302, + 319, + 526, + 413 + ], + "type": "text", + "content": "The results show that DINT Transformer consistently outperforms DIFF Transformer across all datasets and varying numbers of demonstration samples. The improvement in average accuracy is substantial, with DINT achieving " + }, + { + "bbox": [ + 302, + 319, + 526, + 413 + ], + "type": "inline_equation", + "content": "2.8\\%" + }, + { + "bbox": [ + 302, + 319, + 526, + 413 + ], + "type": "text", + "content": " higher accuracy on TREC, " + }, + { + "bbox": [ + 302, + 319, + 526, + 413 + ], + "type": "inline_equation", + "content": "4.1\\%" + }, + { + "bbox": [ + 302, + 319, + 526, + 413 + ], + "type": "text", + "content": " on TREC-Fine, " + }, + { + "bbox": [ + 302, + 319, + 526, + 413 + ], + "type": "inline_equation", + "content": "4.3\\%" + }, + { + "bbox": [ + 302, + 319, + 526, + 413 + ], + "type": "text", + "content": " on Banking-77, and " + }, + { + "bbox": [ + 302, + 319, + 526, + 413 + ], + "type": "inline_equation", + "content": "1.8\\%" + }, + { + "bbox": [ + 302, + 319, + 526, + 413 + ], + "type": "text", + "content": " on Clinic-150." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 302, + 415, + 526, + 672 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 415, + 526, + 672 + ], + "spans": [ + { + "bbox": [ + 302, + 415, + 526, + 672 + ], + "type": "text", + "content": "Robustness of In-Context Learning Figure 6 presents a comparison of the robustness between DIFF Transformer and DINT Transformer in the context of in-context learning. By analyzing how performance varies with different order permutations of the same set of demonstration examples, we find that smaller performance fluctuations reflect greater robustness and a reduced risk of catastrophic degradation. The evaluation protocol remains consistent with the previously outlined methodology. Figure 6 displays the results of this analysis on the TREC dataset. We examine two prompt configurations: randomly shuffled examples and examples arranged by class in an alternating pattern. In both configurations, DINT Transformer consistently shows smaller performance fluctuations compared to DIFF Transformer, demonstrating that our approach enhances robustness in in-context learning tasks." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 303, + 687, + 409, + 699 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 303, + 687, + 409, + 699 + ], + "spans": [ + { + "bbox": [ + 303, + 687, + 409, + 699 + ], + "type": "text", + "content": "3.5 Ablation Studies" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 302, + 708, + 525, + 775 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 708, + 525, + 775 + ], + "spans": [ + { + "bbox": [ + 302, + 708, + 525, + 775 + ], + "type": "text", + "content": "We perform ablation studies using 1.4B-parameter language models, with the same training setup as the 1.4B model in Section 3.2. Both models have 24 layers, with 16 attention heads for Transformer and 8 for DIFF Transformer, each having a head" + } + ] + } + ], + "index": 13 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "spans": [ + { + "bbox": [ + 286, + 780, + 309, + 791 + ], + "type": "text", + "content": "9819" + } + ] + } + ], + "index": 14 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 6 + }, + { + "para_blocks": [ + { + "type": "table", + "bbox": [ + 96, + 68, + 496, + 152 + ], + "blocks": [ + { + "bbox": [ + 96, + 68, + 496, + 152 + ], + "lines": [ + { + "bbox": [ + 96, + 68, + 496, + 152 + ], + "spans": [ + { + "bbox": [ + 96, + 68, + 496, + 152 + ], + "type": "table", + "html": "
ModelAttention to Answer↑Attention Noise↓
0%25%50%75%100%0%25%50%75%100%
Transformer0.030.030.030.070.090.510.540.520.490.49
DIFF0.270.300.310.320.400.010.020.020.020.01
DINT (Ours)0.350.380.400.410.450.010.010.010.010.01
", + "image_path": "ddee8108d4289803fa60ba4a382bbac425f3e7bab1d0decda8c0579c3bfcf583.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "table_body" + } + ], + "index": 0 + }, + { + "type": "table", + "bbox": [ + 120, + 206, + 474, + 363 + ], + "blocks": [ + { + "bbox": [ + 67, + 158, + 525, + 196 + ], + "lines": [ + { + "bbox": [ + 67, + 158, + 525, + 196 + ], + "spans": [ + { + "bbox": [ + 67, + 158, + 525, + 196 + ], + "type": "text", + "content": "Table 5: Attention scores allocated to answer spans and noise context in the key information retrieval task. The target answer is inserted at varying depths within the context. DINT Transformer allocates more attention to relevant information and effectively minimizes attention noise." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "table_caption" + }, + { + "bbox": [ + 120, + 206, + 474, + 363 + ], + "lines": [ + { + "bbox": [ + 120, + 206, + 474, + 363 + ], + "spans": [ + { + "bbox": [ + 120, + 206, + 474, + 363 + ], + "type": "table", + "html": "
Model#HeadsdGNValid. Set↓Fine-Grained Slices
AR-Hit↓Others↓
DIFF81283.0620.8803.247
-GroupNorm81283.1220.9113.309
with λinit = 0.881283.0650.8833.250
with λinit = 0.581283.0660.8823.251
DINT (Ours)81283.0550.8753.243
-GroupNorm81283.0750.8933.256
with λinit = 0.881283.0560.8773.245
with λinit = 0.581283.0580.8783.245
", + "image_path": "ec61c605f3fd8f1c2df2ac490a8169e3fe39507a6426c20c2d19b7df1e53d7ed.jpg" + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "table_body" + } + ], + "index": 2 + }, + { + "bbox": [ + 199, + 370, + 392, + 383 + ], + "lines": [ + { + "bbox": [ + 199, + 370, + 392, + 383 + ], + "spans": [ + { + "bbox": [ + 199, + 370, + 392, + 383 + ], + "type": "text", + "content": "Table 6: Ablation Studies of 1.4B-Size Models." + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "text" + }, + { + "bbox": [ + 67, + 405, + 151, + 417 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 405, + 151, + 417 + ], + "spans": [ + { + "bbox": [ + 67, + 405, + 151, + 417 + ], + "type": "text", + "content": "dimension of 128." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 67, + 428, + 291, + 508 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 428, + 291, + 508 + ], + "spans": [ + { + "bbox": [ + 67, + 428, + 291, + 508 + ], + "type": "text", + "content": "Table 6 reports the fine-grained loss on the validation set, breaking it into two components: \"AR-Hit\" and \"Others.\" \"AR-Hit\" evaluates the model's ability to recall previously seen n-grams, while \"Others\" represents tokens that are either frequent or not recalled from the context." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 67, + 518, + 291, + 774 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 518, + 291, + 774 + ], + "spans": [ + { + "bbox": [ + 67, + 518, + 291, + 774 + ], + "type": "text", + "content": "As shown in Table 6, we performed ablation studies on various design choices in DINT Transformer and compared them with Transformer variants. All models are of similar size and training FLOPs for a fair comparison. The results indicate that our method outperforms DIFF Transformer in both overall loss and fine-grained loss. When GroupNorm is removed, the performance of DIFF Transformer is significantly affected, while DINT Transformer shows a smaller impact. This is because we ensure the row normalization of the attention matrix, which improves the model's overall robustness. Additionally, when using constant initialization for lambda, we observe a slight decrease in performance, but the model still maintains a high level of performance. This demonstrates the effectiveness of our initialization method and shows that the model is robust to different initialization choices." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 303, + 404, + 387, + 417 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 303, + 404, + 387, + 417 + ], + "spans": [ + { + "bbox": [ + 303, + 404, + 387, + 417 + ], + "type": "text", + "content": "4 Conclusions" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 302, + 435, + 527, + 558 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 435, + 527, + 558 + ], + "spans": [ + { + "bbox": [ + 302, + 435, + 527, + 558 + ], + "type": "text", + "content": "We propose DINT Transformer, which integrates global attention statistics into DIFF Transformer to reduce noise and enhance focus on key words. This improves the model's ability to capture global information, ensuring better stability and scalability. Experiments show DINT Transformer excels in long-sequence modeling, key information retrieval, and in-context learning, making it highly promising for NLP tasks requiring global context awareness." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 303, + 581, + 384, + 594 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 303, + 581, + 384, + 594 + ], + "spans": [ + { + "bbox": [ + 303, + 581, + 384, + 594 + ], + "type": "text", + "content": "5 Limitations" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 302, + 613, + 526, + 761 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 302, + 613, + 526, + 761 + ], + "spans": [ + { + "bbox": [ + 302, + 613, + 526, + 761 + ], + "type": "text", + "content": "While the integration mechanism in DINT Transformer has significantly improved model performance, this design inevitably introduces additional computational complexity. These computational characteristics present new optimization opportunities for large-scale model deployment, particularly when processing long-sequence inputs. Through our algorithm-system co-design approach, we are actively developing more efficient implementations to further enhance the computational efficiency of DINT Transformer." + } + ] + } + ], + "index": 10 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 286, + 780, + 310, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 286, + 780, + 310, + 791 + ], + "spans": [ + { + "bbox": [ + 286, + 780, + 310, + 791 + ], + "type": "text", + "content": "9820" + } + ] + } + ], + "index": 11 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 7 + }, + { + "para_blocks": [ + { + "bbox": [ + 69, + 71, + 127, + 83 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 71, + 127, + 83 + ], + "spans": [ + { + "bbox": [ + 69, + 71, + 127, + 83 + ], + "type": "text", + "content": "References" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 69, + 90, + 291, + 774 + ], + "type": "list", + "angle": 0, + "index": 13, + "blocks": [ + { + "bbox": [ + 69, + 90, + 291, + 135 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 90, + 291, + 135 + ], + "spans": [ + { + "bbox": [ + 69, + 90, + 291, + 135 + ], + "type": "text", + "content": "Amanda Bertsch, Maor Ivgi, Uri Alon, Jonathan Berant, Matthew R. Gormley, and Graham Neubig. 2024. In-context learning with long-context models: An in-depth exploration. arXiv preprint, arXiv:2405.00200." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 69, + 143, + 290, + 200 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 143, + 290, + 200 + ], + "spans": [ + { + "bbox": [ + 69, + 143, + 290, + 200 + ], + "type": "text", + "content": "Inigo Casanueva, Tadas Temcinas, Daniela Gerz, Matthew Henderson, and Ivan Vulic. 2020. Efficient intent detection with dual sentence encoders. In Proceedings of the 2nd Workshop on Natural Language Processing for Conversational AI, pages 38-45." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 207, + 290, + 263 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 207, + 290, + 263 + ], + "spans": [ + { + "bbox": [ + 69, + 207, + 290, + 263 + ], + "type": "text", + "content": "L Gao, J Tow, B Abbasi, S Biderman, S Black, A DiPofi, C Foster, L Golding, J Hsu, A Le Noac'h, and 1 others. 2023. A framework for few-shot language model evaluation, 12 2023. URL https://zenodo.org/records/10256836, 7." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 69, + 271, + 290, + 305 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 271, + 290, + 305 + ], + "spans": [ + { + "bbox": [ + 69, + 271, + 290, + 305 + ], + "type": "text", + "content": "Xinyang Geng and Hao Liu. 2023. Openllama: An open reproduction of llama. URL: https://github.com/openlm-research/open_llama." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 69, + 313, + 290, + 370 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 313, + 290, + 370 + ], + "spans": [ + { + "bbox": [ + 69, + 313, + 290, + 370 + ], + "type": "text", + "content": "Eduard Hovy, Laurie Gerber, Ulf Hermjakob, ChinYew Lin, and Deepak Ravichandran. 2001. Toward semantics-based answer pinpointing. In Proceedings of the first international conference on Human language technology research." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 69, + 378, + 290, + 411 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 378, + 290, + 411 + ], + "spans": [ + { + "bbox": [ + 69, + 378, + 290, + 411 + ], + "type": "text", + "content": "Greg Kamradt. 2023. Needle in a haystack - pressure testing llms. https://github.com/gkamradt/LLMTest_NeedleInAHaystack/tree/main." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 69, + 420, + 291, + 529 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 420, + 291, + 529 + ], + "spans": [ + { + "bbox": [ + 69, + 420, + 291, + 529 + ], + "type": "text", + "content": "Stefan Larson, Anish Mahendran, Joseph J Peper, Christopher Clarke, Andrew Lee, Parker Hill, Jonathan K Kummerfeld, Kevin Leach, Michael A Laurenzano, Lingjia Tang, and et al. 2019. An evaluation dataset for intent classification and out-of-scope prediction. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 1311-1316." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 69, + 539, + 290, + 582 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 539, + 290, + 582 + ], + "spans": [ + { + "bbox": [ + 69, + 539, + 290, + 582 + ], + "type": "text", + "content": "Hao Liu, Wilson Yan, Matei Zaharia, and Pieter Abbeel. 2024. World model on million-length video and language with ringattention. arXiv preprint, arXiv:2402.08268." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 69, + 592, + 290, + 646 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 592, + 290, + 646 + ], + "spans": [ + { + "bbox": [ + 69, + 592, + 290, + 646 + ], + "type": "text", + "content": "Yao Lu, Max Bartolo, Alastair Moore, Sebastian Riedel, and Pontus Stenetorp. 2021. Fantastically ordered prompts and where to find them: Overcoming few-shot prompt order sensitivity. arXiv preprint arXiv:2104.08786." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 69, + 655, + 290, + 700 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 655, + 290, + 700 + ], + "spans": [ + { + "bbox": [ + 69, + 655, + 290, + 700 + ], + "type": "text", + "content": "Zhen Qin, Xiaodong Han, Weixuan Sun, Dongxu Li, Lingpeng Kong, Nick Barnes, and Yiran Zhong. 2022. The devil in linear transformer. arXiv preprint arXiv:2210.10340." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 69, + 709, + 290, + 743 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 709, + 290, + 743 + ], + "spans": [ + { + "bbox": [ + 69, + 709, + 290, + 743 + ], + "type": "text", + "content": "Prajit Ramachandran, Barret Zoph, and Quoc V Le. 2017. Swish: a self-gated activation function. arXiv preprint arXiv:1710.05941, 7(1):5." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 69, + 751, + 290, + 774 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 751, + 290, + 774 + ], + "spans": [ + { + "bbox": [ + 69, + 751, + 290, + 774 + ], + "type": "text", + "content": "Nir Ratner, Yoav Levine, Yonatan Belinkov, Ori Ram, Inbal Magar, Omri Abend, Ehud Karpas, Amnon" + } + ] + } + ], + "index": 12 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 304, + 72, + 525, + 633 + ], + "type": "list", + "angle": 0, + "index": 25, + "blocks": [ + { + "bbox": [ + 314, + 72, + 525, + 127 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 314, + 72, + 525, + 127 + ], + "spans": [ + { + "bbox": [ + 314, + 72, + 525, + 127 + ], + "type": "text", + "content": "Shashua, Kevin Leyton-Brown, and Yoav Shoham. 2023. Parallel context windows for large language models. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics, pages 6383-6402." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 304, + 136, + 525, + 212 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 136, + 525, + 212 + ], + "spans": [ + { + "bbox": [ + 304, + 136, + 525, + 212 + ], + "type": "text", + "content": "Machel Reid, Nikolay Savinov, Denis Teptyashin, Dmitry Lepikhin, Timothy Lillicrap, Jean-Baptiste Alayrac, Radu Soricut, Angeliki Lazaridou, Orhan Firat, Julian Schrittwieser, and et al. 2024. Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. arXiv preprint, arXiv:2403.05530." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 304, + 222, + 525, + 244 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 222, + 525, + 244 + ], + "spans": [ + { + "bbox": [ + 304, + 222, + 525, + 244 + ], + "type": "text", + "content": "Noam Shazeer. 2020. Glu variants improve transformer. arXiv preprint arXiv:2002.05202." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 304, + 252, + 525, + 318 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 252, + 525, + 318 + ], + "spans": [ + { + "bbox": [ + 304, + 252, + 525, + 318 + ], + "type": "text", + "content": "Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothee Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, and 1 others. 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971." + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 304, + 327, + 525, + 360 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 327, + 525, + 360 + ], + "spans": [ + { + "bbox": [ + 304, + 327, + 525, + 360 + ], + "type": "text", + "content": "Jonathan Tow. 2023. Stablelm alpha v2 models. https://huggingface.co/stabilityai/stablelm-base-alpha-3b-v2." + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 304, + 369, + 525, + 401 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 369, + 525, + 401 + ], + "spans": [ + { + "bbox": [ + 304, + 369, + 525, + 401 + ], + "type": "text", + "content": "Jonathan Tow, Marco Bellagente, Dakota Mahan, and Carlos Riquelme. 2023. Stablelm 3b 4e1t. https://aka.ms/StableLM-3B-4E1T." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 304, + 410, + 525, + 433 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 410, + 525, + 433 + ], + "spans": [ + { + "bbox": [ + 304, + 410, + 525, + 433 + ], + "type": "text", + "content": "A Vaswani. 2017. Attention is all you need. Advances in Neural Information Processing Systems." + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 304, + 441, + 525, + 507 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 441, + 525, + 507 + ], + "spans": [ + { + "bbox": [ + 304, + 441, + 525, + 507 + ], + "type": "text", + "content": "Hongyu Wang, Shuming Ma, Shaohan Huang, Li Dong, Wenhui Wang, Zhiliang Peng, Yu Wu, Payal Bajaj, Saksham Singhal, Alon Benhaim, and 1 others. 2023. Magneto: a foundation transformer. In International Conference on Machine Learning, pages 36077-36092. PMLR." + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 304, + 516, + 525, + 550 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 516, + 525, + 550 + ], + "spans": [ + { + "bbox": [ + 304, + 516, + 525, + 550 + ], + "type": "text", + "content": "Yuxin Wu and Kaiming He. 2018. Group normalization. In Proceedings of the European conference on computer vision (ECCV), pages 3-19." + } + ] + } + ], + "index": 22 + }, + { + "bbox": [ + 304, + 557, + 525, + 592 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 557, + 525, + 592 + ], + "spans": [ + { + "bbox": [ + 304, + 557, + 525, + 592 + ], + "type": "text", + "content": "Tianzhu Ye, Li Dong, Yuqing Xia, Yutao Sun, Yi Zhu, Gao Huang, and Furu Wei. 2024. Differential transformer. arXiv preprint arXiv:2410.05258." + } + ] + } + ], + "index": 23 + }, + { + "bbox": [ + 304, + 599, + 525, + 633 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 304, + 599, + 525, + 633 + ], + "spans": [ + { + "bbox": [ + 304, + 599, + 525, + 633 + ], + "type": "text", + "content": "Biao Zhang and Rico Sennrich. 2019. Root mean square layer normalization. Advances in Neural Information Processing Systems, 32." + } + ] + } + ], + "index": 24 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 286, + 781, + 308, + 791 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 286, + 781, + 308, + 791 + ], + "spans": [ + { + "bbox": [ + 286, + 781, + 308, + 791 + ], + "type": "text", + "content": "9821" + } + ] + } + ], + "index": 26 + } + ], + "page_size": [ + 595, + 841 + ], + "page_idx": 8 + } + ], + "_backend": "vlm", + "_version_name": "2.6.4" +} \ No newline at end of file