Saurabh502 commited on
Commit
36bd29a
·
verified ·
1 Parent(s): 04c01c9

Update index.html

Browse files
Files changed (1) hide show
  1. index.html +240 -86
index.html CHANGED
@@ -20,173 +20,212 @@
20
  justify-content: center;
21
  align-items: center;
22
  min-height: 100vh;
23
- background-image: url('quiz-background.jpg'); /* Replace with your image */
24
  background-size: cover;
25
  background-position: center;
26
  }
27
 
28
  #root {
29
- background-color: rgba(255, 255, 255, 0.95); /* Slightly transparent white */
30
  padding: 30px;
31
  border-radius: 12px;
32
- box-shadow: 0 8px 20px rgba(0, 0, 0, 0.1); /* Softer shadow */
33
- width: 80%; /* Slightly wider */
34
- max-width: 800px; /* Maximum width for larger screens */
35
  text-align: center;
36
- box-sizing: border-box; /* Include padding in width */
37
  }
38
 
39
  h1 {
40
  font-size: 2rem;
41
  margin-bottom: 20px;
42
- color: #3498db; /* A brighter, more modern blue */
43
- font-weight: 600; /* Semi-bold for headings */
44
  }
45
 
46
  p {
47
  font-size: 1.1rem;
48
- margin-bottom: 25px; /* Increased margin for better spacing */
49
- color: #555; /* Darker, more readable gray */
50
- line-height: 1.7; /* Improved line height for readability */
51
  }
52
 
53
  #question-area {
54
  font-size: 1.2rem;
55
  margin-bottom: 20px;
56
  padding: 15px;
57
- background-color: #e8f0fa; /* Very light blue for question area */
58
  border-radius: 8px;
59
- border: 1px solid #b8c6da; /* Subtle border */
60
- color: #2c3e50; /* Dark blue for question text */
61
- text-align: left; /* Left-align the question */
62
  }
63
 
64
  #answer-options {
65
  display: flex;
66
  flex-direction: column;
67
- align-items: stretch; /* Stretch options to container width */
68
- margin-bottom: 25px; /* Increased margin */
69
  }
70
 
71
  .answer-option {
72
  padding: 15px;
73
- margin-bottom: 12px; /* Increased margin */
74
- background-color: #fff; /* White background for options */
75
  border-radius: 8px;
76
- border: 1px solid #ddd; /* Lighter border */
77
  cursor: pointer;
78
- transition: background-color 0.3s ease, transform 0.2s ease; /* Smooth transition */
79
- font-size: 1.1rem; /* Slightly larger font */
80
- text-align: left; /* Left-align the options text */
81
- box-shadow: 0 2px 4px rgba(0, 0, 0, 0.05); /* Very subtle shadow */
82
  }
83
 
84
  .answer-option:hover {
85
- background-color: #f0f8ff; /* Very light blue on hover */
86
- transform: translateY(-2px); /* Slight lift on hover */
87
- border-color: #a6d4fa; /* Slightly more prominent border on hover */
88
  }
89
 
90
  .answer-option.selected {
91
- background-color: #a6d4fa; /* Light blue for selected option */
92
- border-color: #3498db; /* Stronger blue border for selected */
93
- color: #fff; /* White text for selected option */
94
- box-shadow: 0 4px 6px rgba(0, 0, 0, 0.1); /* Slightly stronger shadow */
95
  }
96
 
97
  .answer-option.correct {
98
- background-color: #86ef7d; /* Light green for correct answer */
99
- border-color: #22c55e; /* Stronger green border */
100
  color: #fff;
101
- font-weight: 600; /* Bold correct answer */
102
  }
103
 
104
  .answer-option.incorrect {
105
- background-color: #fca5a5; /* Light red for incorrect answer */
106
- border-color: #dc2626; /* Stronger red border */
107
  color: #fff;
108
- font-weight: 600; /* Bold incorrect answer */
109
  }
110
 
111
  #result-message {
112
  font-size: 1.2rem;
113
  margin-bottom: 25px;
114
- font-weight: 500; /* Medium font weight for message */
115
  }
116
 
117
  .correct-message {
118
- color: #22c55e; /* Strong green for correct message */
119
  }
120
 
121
  .incorrect-message {
122
- color: #dc2626; /* Strong red for incorrect message */
123
  }
124
 
125
-
126
- #next-button, #skip-button, #reset-button {
127
  padding: 12px 25px;
128
  font-size: 1.1rem;
129
- margin: 5px 10px; /* Added horizontal margin between buttons */
130
  border-radius: 8px;
131
  border: none;
132
  cursor: pointer;
133
- transition: background-color 0.3s ease, transform 0.2s ease, box-shadow 0.3s ease; /* Added box-shadow transition */
134
- font-weight: 500; /* Medium font weight for buttons */
135
- box-shadow: 0 2px 5px rgba(0, 0, 0, 0.1); /* Subtle shadow for buttons */
136
  }
137
 
138
  #next-button {
139
- background-color: #3498db; /* Blue for next button */
140
  color: #fff;
141
  }
142
  #next-button:hover {
143
  background-color: #2980b9;
144
  transform: translateY(-2px);
145
- box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15); /* Increased shadow on hover */
146
  }
147
  #next-button:disabled {
148
- background-color: #b8c6da; /* Gray for disabled next button */
149
  cursor: not-allowed;
150
- transform: none; /* Remove transform on disabled */
151
- box-shadow: none; /* Remove shadow on disabled */
152
  }
153
 
154
  #skip-button {
155
- background-color: #f39c12; /* Orange for skip button */
156
  color: #fff;
157
  }
158
  #skip-button:hover {
159
  background-color: #e67e22;
160
  transform: translateY(-2px);
161
- box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15); /* Increased shadow on hover */
162
  }
163
 
164
  #reset-button {
165
- background-color: #2ecc71; /* Green for reset button */
166
  color: #fff;
167
  }
168
  #reset-button:hover {
169
  background-color: #27ae60;
170
  transform: translateY(-2px);
171
- box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15); /* Increased shadow on hover */
 
 
 
 
 
 
 
 
 
 
172
  }
173
 
174
  #final-score-area {
175
  font-size: 1.5rem;
176
  font-weight: 600;
177
  margin-bottom: 30px;
178
- color: #2c3e50; /* Dark blue for final score */
179
  }
180
 
181
  .pass-message {
182
- color: #22c55e; /* Green for pass message */
183
  }
184
 
185
  .fail-message {
186
- color: #dc2626; /* Red for fail message */
187
  }
188
 
189
- @media (max-width: 768px) { /* Styles for tablets and smaller screens */
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
190
  #root {
191
  width: 95%;
192
  padding: 20px;
@@ -200,7 +239,7 @@
200
  font-size: 1.1rem;
201
  padding: 12px;
202
  }
203
- #next-button, #skip-button, #reset-button {
204
  padding: 10px 20px;
205
  font-size: 1rem;
206
  }
@@ -212,12 +251,12 @@
212
  }
213
  }
214
 
215
- @media (max-width: 480px) { /* Styles for phones */
216
  #root {
217
  width: 100%;
218
  padding: 15px;
219
- border-radius: 0; /* Remove border radius on phones */
220
- box-shadow: none; /* Remove shadow on phones */
221
  }
222
  .answer-option {
223
  padding: 10px;
@@ -228,10 +267,10 @@
228
  font-size: 1rem;
229
  padding: 10px;
230
  }
231
- #next-button, #skip-button, #reset-button {
232
  padding: 10px 18px;
233
  font-size: 0.9rem;
234
- margin: 5px 5px; /* Reduced horizontal margin on phones */
235
  }
236
  h1 {
237
  font-size: 1.5rem;
@@ -262,6 +301,7 @@
262
  "To predict stock market trends.",
263
  ],
264
  correctAnswer: "To understand and generate human language.",
 
265
  },
266
  {
267
  question: "Which of the following is NOT a stage in the development of language models?",
@@ -272,6 +312,7 @@
272
  "Quantum language models (QLM)",
273
  ],
274
  correctAnswer: "Quantum language models (QLM)",
 
275
  },
276
  {
277
  question: "What is the basic idea behind statistical language models (SLMs)?",
@@ -282,6 +323,7 @@
282
  "To pre-train Transformer models.",
283
  ],
284
  correctAnswer: "To build word prediction models based on the Markov assumption.",
 
285
  },
286
  {
287
  question: "What is a limitation of SLMs?",
@@ -292,6 +334,7 @@
292
  "They are not used in NLP.",
293
  ],
294
  correctAnswer: "They suffer from the curse of dimensionality.",
 
295
  },
296
  {
297
  question: "What is the main characteristic of neural language models (NLMs)?",
@@ -302,6 +345,7 @@
302
  "They use n-gram models.",
303
  ],
304
  correctAnswer: "They characterize the probability of word sequences by neural networks.",
 
305
  },
306
  {
307
  question: "What concept did the work in [1] introduce?",
@@ -312,6 +356,7 @@
312
  "Pre-training and fine-tuning.",
313
  ],
314
  correctAnswer: "Distributed representation of words.",
 
315
  },
316
  {
317
  question: "What is word2vec?",
@@ -322,16 +367,19 @@
322
  "A pre-trained language model.",
323
  ],
324
  correctAnswer: "A simplified shallow neural network for learning distributed word representations.",
 
325
  },
326
  {
327
  question: "What was an early attempt at capturing context-aware word representations?",
328
  options: ["BERT", "ELMo", "GPT-2", "word2vec"],
329
  correctAnswer: "ELMo",
 
330
  },
331
  {
332
  question: "Which architecture is highly parallelizable and used in BERT?",
333
  options: ["RNN", "LSTM", "Transformer", "MLP"],
334
  correctAnswer: "Transformer",
 
335
  },
336
  {
337
  question: "What is a key feature of BERT?",
@@ -342,6 +390,7 @@
342
  "It predicts the next word based on the most recent context.",
343
  ],
344
  correctAnswer: "It pre-trains bidirectional language models.",
 
345
  },
346
  {
347
  question: "What paradigm did BERT inspire?",
@@ -352,6 +401,7 @@
352
  "Distributed representation learning.",
353
  ],
354
  correctAnswer: "Pre-training and fine-tuning.",
 
355
  },
356
  {
357
  question: "What do researchers find about scaling PLMs?",
@@ -362,6 +412,7 @@
362
  "It only works for small models.",
363
  ],
364
  correctAnswer: "It often leads to an improved model capacity.",
 
365
  },
366
  {
367
  question: "What is a characteristic of large-sized PLMs (LLMs)?",
@@ -372,6 +423,7 @@
372
  "They are less complex than SLMs.",
373
  ],
374
  correctAnswer: "They display different behaviors from smaller PLMs.",
 
375
  },
376
  {
377
  question: "What is an example of an emergent ability in LLMs?",
@@ -382,8 +434,9 @@
382
  "Using only n-gram models.",
383
  ],
384
  correctAnswer: "Solving few-shot tasks through in-context learning.",
 
385
  },
386
- {
387
  question: "What does the term 'LLM' stand for?",
388
  options: [
389
  "Limited Language Model",
@@ -392,11 +445,13 @@
392
  "Logical Language Model",
393
  ],
394
  correctAnswer: "Large Language Model",
 
395
  },
396
  {
397
  question: "Which of the following is a remarkable application of LLMs?",
398
  options: ["ELMo", "ChatGPT", "word2vec", "SLM"],
399
  correctAnswer: "ChatGPT",
 
400
  },
401
  {
402
  question: "What is a key focus of the latest language models (e.g., GPT-4)?",
@@ -407,6 +462,7 @@
407
  "Using only statistical methods.",
408
  ],
409
  correctAnswer: "Complex task solving.",
 
410
  },
411
  {
412
  question: "How do LLMs differ from small PLMs in accessing them?",
@@ -417,6 +473,7 @@
417
  "They are accessed using n-gram models.",
418
  ],
419
  correctAnswer: "They are accessed through the prompting interface.",
 
420
  },
421
  {
422
  question: "What is a challenge in developing LLMs?",
@@ -427,6 +484,7 @@
427
  "It is very costly to train them due to huge demand for computation resources.",
428
  ],
429
  correctAnswer: "It is very costly to train them due to huge demand for computation resources.",
 
430
  },
431
  {
432
  question: "What is a potential issue with LLMs despite their capacities?",
@@ -437,6 +495,7 @@
437
  "They do not require effective control approaches.",
438
  ],
439
  correctAnswer: "They are likely to produce toxic, fictitious, or harmful content.",
 
440
  },
441
  {
442
  question: "What are the four major aspects of LLMs covered in the survey?",
@@ -447,6 +506,7 @@
447
  "Input, processing, output, and storage.",
448
  ],
449
  correctAnswer: "Pre-training, adaptation, utilization, and capacity evaluation.",
 
450
  },
451
  {
452
  question: "What is the typical parameter size of LLMs?",
@@ -457,6 +517,7 @@
457
  "Less than 100 parameters.",
458
  ],
459
  correctAnswer: "Billions or hundreds of billions of parameters.",
 
460
  },
461
  {
462
  question: "On what type of data are LLMs typically trained?",
@@ -467,6 +528,7 @@
467
  "Audio data only.",
468
  ],
469
  correctAnswer: "Massive text data.",
 
470
  },
471
  {
472
  question: "What strong capacity do LLMs exhibit?",
@@ -477,6 +539,7 @@
477
  "Performing only simple calculations.",
478
  ],
479
  correctAnswer: "Solving complex tasks via text generation.",
 
480
  },
481
  {
482
  question: "What does the survey introduce as basic background for LLMs?",
@@ -487,6 +550,7 @@
487
  "Only scaling laws.",
488
  ],
489
  correctAnswer: "Scaling laws, emergent abilities, and key techniques.",
 
490
  },
491
  {
492
  question: "What is the key to understanding the development of language models in research history?",
@@ -497,6 +561,7 @@
497
  "Ignoring the evolution of model capacities.",
498
  ],
499
  correctAnswer: "The leap from language modeling to task solving.",
 
500
  },
501
  {
502
  question: "Which of the following is NOT a characteristic of LLMs?",
@@ -507,6 +572,7 @@
507
  "Using Transformer language models.",
508
  ],
509
  correctAnswer: "Training on small datasets.",
 
510
  },
511
  {
512
  question: "What is a major difference between LLMs and previous smaller PLMs?",
@@ -517,6 +583,7 @@
517
  "LLMs do not require large-scale data.",
518
  ],
519
  correctAnswer: "LLMs display surprising emergent abilities.",
 
520
  },
521
  {
522
  question: "How has the research paradigm shifted towards the use of LLMs in NLP?",
@@ -527,6 +594,7 @@
527
  "Ignoring pre-training and fine-tuning.",
528
  ],
529
  correctAnswer: "LLMs serve as a general-purpose language task solver.",
 
530
  },
531
  {
532
  question: "What is a challenge that traditional search engines face with the advent of LLMs?",
@@ -537,6 +605,7 @@
537
  "Easier data processing.",
538
  ],
539
  correctAnswer: "New information seeking way through AI chatbots.",
 
540
  },
541
  {
542
  question: "What is a current trend in computer vision (CV) research related to LLMs?",
@@ -547,6 +616,7 @@
547
  "Focusing solely on image classification.",
548
  ],
549
  correctAnswer: "Developing ChatGPT-like vision-language models.",
 
550
  },
551
  {
552
  question: "What is a potential impact of the new wave of LLM technology?",
@@ -557,6 +627,7 @@
557
  "Slower development of AI algorithms.",
558
  ],
559
  correctAnswer: "A prosperous ecosystem of real-world applications based on LLMs.",
 
560
  },
561
  {
562
  question: "What is one of the mysterious aspects of LLMs?",
@@ -567,6 +638,7 @@
567
  "Why they only work on small datasets.",
568
  ],
569
  correctAnswer: "Why emergent abilities occur in LLMs.",
 
570
  },
571
  {
572
  question: "Why is it difficult for the research community to train capable LLMs?",
@@ -577,6 +649,7 @@
577
  "Because LLMs are mainly trained by academia.",
578
  ],
579
  correctAnswer: "Because it is very costly to carry out repetitive studies.",
 
580
  },
581
  {
582
  question: "What is a challenge in aligning LLMs?",
@@ -587,6 +660,7 @@
587
  "Reducing their computational costs.",
588
  ],
589
  correctAnswer: "Aligning them with human values or preferences.",
 
590
  },
591
  {
592
  question: "What does the paper 'Planning for AGI and beyond' discuss?",
@@ -597,6 +671,7 @@
597
  "Only short-term plans for AI development.",
598
  ],
599
  correctAnswer: "Short-term and long-term plans to approach AGI.",
 
600
  },
601
  {
602
  question: "What is a recent argument about GPT-4?",
@@ -607,6 +682,7 @@
607
  "It can only perform simple tasks.",
608
  ],
609
  correctAnswer: "It might be considered an early version of an AGI system.",
 
610
  },
611
  {
612
  question: "How is Microsoft 365 being empowered?",
@@ -617,32 +693,37 @@
617
  "By using only statistical models.",
618
  ],
619
  correctAnswer: "By LLMs to automate office work.",
 
620
  },
621
  {
622
- question: "What is a suggestion for choosing layer normalization in LLMs?",
623
- options: ["Post RMSNorm", "Pre RMSNorm", "Post LN", "No Normalization"],
624
- correctAnswer: "Pre RMSNorm",
 
625
  },
626
  {
627
- question: "Which activation function is recommended for stronger generalization and training stability?",
628
- options: ["ReLU", "Sigmoid", "SwiGLU", "Tanh"],
629
- correctAnswer: "SwiGLU",
 
630
  },
631
  {
632
- question: "Which position embedding is considered a better choice for LLMs?",
633
- options: ["Absolute Positional Embedding", "Relative Positional Embedding", "RoPE", "Sinusoidal Positional Encoding"],
634
- correctAnswer: "RoPE",
 
635
  },
636
  {
637
  question: "What is the primary role of pre-training in LLMs?",
638
- options: [
639
  "To fine-tune models for specific tasks.",
640
  "To encode general knowledge from large-scale corpus.",
641
  "To reduce the size of the model.",
642
  "To improve inference speed.",
643
  ],
644
  correctAnswer: "To encode general knowledge from large-scale corpus.",
645
- },
 
646
  {
647
  question: "What type of data is included in the arXiv Dataset?",
648
  options: [
@@ -652,11 +733,13 @@ options: [
652
  "Image and video data.",
653
  ],
654
  correctAnswer: "Scientific publication data.",
 
655
  },
656
  {
657
  question: "What is the approximate size of the peS2o dataset?",
658
  options: ["42MB", "42GB", "42TB", "42B tokens"],
659
  correctAnswer: "42B tokens",
 
660
  },
661
  {
662
  question: "What is a characteristic of the articles in Wikipedia?",
@@ -667,11 +750,13 @@ options: [
667
  "They cover only a narrow range of topics.",
668
  ],
669
  correctAnswer: "They are composed in an expository style with references.",
 
670
  },
671
  {
672
  question: "What is a technique used to improve memory efficiency and throughput of deployed LLMs?",
673
  options: ["Data Parallelism", "Tensor Parallelism", "Pipeline Parallelism", "PagedAttention"],
674
  correctAnswer: "PagedAttention",
 
675
  },
676
  {
677
  question: "How does PagedAttention partition sequences?",
@@ -682,6 +767,7 @@ options: [
682
  "Into overlapping segments.",
683
  ],
684
  correctAnswer: "Into subsequences.",
 
685
  },
686
  {
687
  question: "What is the benefit of using PagedAttention?",
@@ -692,11 +778,13 @@ options: [
692
  "Improves model accuracy.",
693
  ],
694
  correctAnswer: "Increases GPU utilization and enables efficient memory sharing.",
 
695
  },
696
  {
697
  question: "What type of floating-point number was predominantly used for pre-training in previous PLMs like BERT?",
698
  options: ["FP16", "BF16", "FP32", "INT8"],
699
  correctAnswer: "FP32",
 
700
  },
701
  {
702
  question: "Why have some studies started to use FP16 for pre-training LLMs?",
@@ -707,6 +795,7 @@ options: [
707
  "To avoid the loss of computational accuracy.",
708
  ],
709
  correctAnswer: "To reduce memory usage and communication overhead.",
 
710
  },
711
  {
712
  question: "What is a potential issue with using FP16 for training?",
@@ -717,6 +806,7 @@ options: [
717
  "Improved model performance.",
718
  ],
719
  correctAnswer: "Loss of computational accuracy.",
 
720
  },
721
  {
722
  question: "What is BF16?",
@@ -727,6 +817,7 @@ options: [
727
  "A method for parallel training.",
728
  ],
729
  correctAnswer: "Brain Floating Point, an alternative to FP16.",
 
730
  },
731
  {
732
  question: "How does BF16 compare to FP16 in terms of representation accuracy for pre-training?",
@@ -737,6 +828,7 @@ options: [
737
  "BF16 is not suitable for pre-training.",
738
  ],
739
  correctAnswer: "BF16 generally performs better than FP16.",
 
740
  },
741
  {
742
  question: "Which training technique is often used jointly with 3D parallelism to improve training throughput?",
@@ -747,6 +839,7 @@ options: [
747
  "PagedAttention.",
748
  ],
749
  correctAnswer: "Mixed precision training.",
 
750
  },
751
  {
752
  question: "What type of parallelism was used to train BLOOM on 384 A100 GPUs?",
@@ -757,6 +850,7 @@ options: [
757
  "Only pipeline parallelism.",
758
  ],
759
  correctAnswer: "8-way data parallelism, 4-way tensor parallelism, and 12-way pipeline parallelism.",
 
760
  },
761
  {
762
  question: "What is the primary approach to using LLMs after pre-training or adaptation tuning?",
@@ -767,11 +861,13 @@ options: [
767
  "Using only n-gram models.",
768
  ],
769
  correctAnswer: "Designing suitable prompting strategies.",
 
770
  },
771
  {
772
  question: "What is a representative prompting method discussed in the text?",
773
  options: ["Fine-tuning", "Backpropagation", "In-context learning", "Random search"],
774
  correctAnswer: "In-context learning",
 
775
  },
776
  {
777
  question: "What does in-context learning involve?",
@@ -782,11 +878,13 @@ options: [
782
  "Employing only manual creation of prompts.",
783
  ],
784
  correctAnswer: "Formulating task description and demonstrations in natural language text.",
 
785
  },
786
  {
787
  question: "What is the process of manually creating a suitable prompt also called?",
788
  options: ["Automatic prompt optimization", "Prompt engineering", "Prompt tuning", "Prompt generation"],
789
  correctAnswer: "Prompt engineering",
 
790
  },
791
  {
792
  question: "What is the impact of a well-designed prompt on LLMs?",
@@ -797,16 +895,19 @@ options: [
797
  "It makes LLMs generate random outputs.",
798
  ],
799
  correctAnswer: "It is very helpful to elicit the abilities of LLMs.",
 
800
  },
801
  {
802
  question: "What type of data is considered well-organized with algorithmic logic and programming flow?",
803
  options: ["Natural language text", "Image data", "Code data", "Audio data"],
804
  correctAnswer: "Code data",
 
805
  },
806
  {
807
  question: "What ability do models trained on code show?",
808
  options: ["Weak reasoning ability", "Strong reasoning ability", "No reasoning ability", "Only language generation ability"],
809
  correctAnswer: "Strong reasoning ability",
 
810
  },
811
  {
812
  question: "What is a hypothesis regarding code data and LLMs' reasoning performance?",
@@ -817,6 +918,7 @@ options: [
817
  "Code data is only useful for code generation.",
818
  ],
819
  correctAnswer: "Code data may be useful to improve the reasoning performance of LLMs.",
 
820
  },
821
  {
822
  question: "What is a characteristic of LLMs' text generation quality?",
@@ -827,6 +929,7 @@ options: [
827
  "It cannot be evaluated.",
828
  ],
829
  correctAnswer: "It is comparable to human-written texts.",
 
830
  },
831
  {
832
  question: "How can LLMs be used in the context of generation evaluation?",
@@ -837,6 +940,7 @@ options: [
837
  "To perform only statistical analysis.",
838
  ],
839
  correctAnswer: "As language generation evaluators.",
 
840
  },
841
  {
842
  question: "What is a limitation of LLMs in specialized generation?",
@@ -847,6 +951,7 @@ options: [
847
  "They are only good at generating code.",
848
  ],
849
  correctAnswer: "They have learned general language patterns but underperform in specialized generation.",
 
850
  },
851
  {
852
  question: "What is a common approach to enhancing LLMs' factual knowledge?",
@@ -857,6 +962,7 @@ options: [
857
  "Ignoring up-to-date information.",
858
  ],
859
  correctAnswer: "Incorporating extracted relevant information into the context.",
 
860
  },
861
  {
862
  question: "What is a finding about smaller models with instruction tuning compared to larger models without it?",
@@ -867,6 +973,7 @@ options: [
867
  "Larger models are always better.",
868
  ],
869
  correctAnswer: "Smaller models can perform better.",
 
870
  },
871
  {
872
  question: "What does instruction tuning enable LLMs to do?",
@@ -877,6 +984,7 @@ options: [
877
  "Perform only without demonstrations.",
878
  ],
879
  correctAnswer: "Follow human instructions to perform specific tasks, even on unseen tasks.",
 
880
  },
881
  {
882
  question: "What is a benefit of instruction tuning?",
@@ -887,6 +995,7 @@ options: [
887
  "It does not improve model performance.",
888
  ],
889
  correctAnswer: "It is much less costly than pre-training.",
 
890
  },
891
  {
892
  question: "What is the primary goal of instruction tuning?",
@@ -897,6 +1006,7 @@ options: [
897
  "To generate random text",
898
  ],
899
  correctAnswer: "To make LLMs better at following instructions",
 
900
  },
901
  {
902
  question: "What kind of tasks does instruction tuning help LLMs perform?",
@@ -907,6 +1017,7 @@ options: [
907
  "No specific tasks",
908
  ],
909
  correctAnswer: "Specific tasks without demonstrations",
 
910
  },
911
  {
912
  question: "How does instruction tuning affect LLMs' ability to follow instructions?",
@@ -917,6 +1028,7 @@ options: [
917
  "It makes them ignore instructions",
918
  ],
919
  correctAnswer: "It enhances their ability",
 
920
  },
921
  {
922
  question: "What have a large number of studies confirmed about instruction tuning?",
@@ -927,6 +1039,7 @@ options: [
927
  "It decreases performance",
928
  ],
929
  correctAnswer: "It achieves superior performance on seen and unseen tasks",
 
930
  },
931
  {
932
  question: "What is a key aspect of high-quality long CoT data curation?",
@@ -937,6 +1050,7 @@ options: [
937
  "Manual data collection only",
938
  ],
939
  correctAnswer: "Using open models or APIs for data synthesis",
 
940
  },
941
  {
942
  question: "What is the basic idea behind creating long CoT response data?",
@@ -947,6 +1061,7 @@ options: [
947
  "Ignoring the prompt structure",
948
  ],
949
  correctAnswer: "Feeding prompts into teacher models",
 
950
  },
951
  {
952
  question: "What is the finding about smaller models with instruction tuning compared to larger models without fine-tuning?",
@@ -957,6 +1072,7 @@ options: [
957
  "Larger models are always better.",
958
  ],
959
  correctAnswer: "Smaller models can perform better.",
 
960
  },
961
  ];
962
 
@@ -968,12 +1084,12 @@ options: [
968
  const [quizEnd, setQuizEnd] = React.useState(false);
969
  const [selectedAnswerColor, setSelectedAnswerColor] = React.useState('');
970
  const [currentQuestionNumber, setCurrentQuestionNumber] = React.useState(1);
971
-
 
972
 
973
  const currentQuestion = quizData[currentQuestionIndex];
974
  const totalQuestions = quizData.length;
975
 
976
-
977
  const handleAnswerSelection = (answer) => {
978
  setSelectedAnswer(answer);
979
  if (answer === currentQuestion.correctAnswer) {
@@ -983,6 +1099,12 @@ options: [
983
  } else {
984
  setMessage('Incorrect!');
985
  setSelectedAnswerColor('incorrect');
 
 
 
 
 
 
986
  }
987
  };
988
 
@@ -1018,6 +1140,12 @@ options: [
1018
  setQuizEnd(false);
1019
  setSelectedAnswerColor('');
1020
  setCurrentQuestionNumber(1);
 
 
 
 
 
 
1021
  };
1022
 
1023
  return (
@@ -1028,13 +1156,13 @@ options: [
1028
  <>
1029
  <div id="question-area">
1030
  Question {currentQuestionNumber}/{totalQuestions}: {currentQuestion.question}
1031
- </div>
1032
  <div id="answer-options">
1033
  {currentQuestion.options.map((option) => (
1034
  <div
1035
  key={option}
1036
  className={`answer-option ${selectedAnswer === option ? selectedAnswerColor : ''} ${selectedAnswer ? 'disabled' : ''} ${selectedAnswerColor && option === currentQuestion.correctAnswer ? 'correct' : ''}`}
1037
- onClick={() => handleAnswerSelection(option)}
1038
  >
1039
  {option}
1040
  </div>
@@ -1054,15 +1182,41 @@ options: [
1054
  <>
1055
  <div id="final-score-area">
1056
  Your Final Score: {score} / {quizData.length} <br/>
1057
- {score >= 40 ? (
1058
  <span className="pass-message">Passed</span>
1059
  ) : (
1060
  <span className="fail-message">Failed</span>
1061
  )}
1062
  </div>
1063
- <button id="reset-button"onClick={handleResetQuiz}>
1064
- Play Again
1065
- </button>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1066
  </>
1067
  )}
1068
  </div>
@@ -1072,4 +1226,4 @@ options: [
1072
  ReactDOM.render(<QuizApp />, document.getElementById("root"));
1073
  </script>
1074
  </body>
1075
- </html>
 
20
  justify-content: center;
21
  align-items: center;
22
  min-height: 100vh;
23
+ background-image: url('quiz-background.jpg');
24
  background-size: cover;
25
  background-position: center;
26
  }
27
 
28
  #root {
29
+ background-color: rgba(255, 255, 255, 0.95);
30
  padding: 30px;
31
  border-radius: 12px;
32
+ box-shadow: 0 8px 20px rgba(0, 0, 0, 0.1);
33
+ width: 80%;
34
+ max-width: 800px;
35
  text-align: center;
36
+ box-sizing: border-box;
37
  }
38
 
39
  h1 {
40
  font-size: 2rem;
41
  margin-bottom: 20px;
42
+ color: #3498db;
43
+ font-weight: 600;
44
  }
45
 
46
  p {
47
  font-size: 1.1rem;
48
+ margin-bottom: 25px;
49
+ color: #555;
50
+ line-height: 1.7;
51
  }
52
 
53
  #question-area {
54
  font-size: 1.2rem;
55
  margin-bottom: 20px;
56
  padding: 15px;
57
+ background-color: #e8f0fa;
58
  border-radius: 8px;
59
+ border: 1px solid #b8c6da;
60
+ color: #2c3e50;
61
+ text-align: left;
62
  }
63
 
64
  #answer-options {
65
  display: flex;
66
  flex-direction: column;
67
+ align-items: stretch;
68
+ margin-bottom: 25px;
69
  }
70
 
71
  .answer-option {
72
  padding: 15px;
73
+ margin-bottom: 12px;
74
+ background-color: #fff;
75
  border-radius: 8px;
76
+ border: 1px solid #ddd;
77
  cursor: pointer;
78
+ transition: background-color 0.3s ease, transform 0.2s ease;
79
+ font-size: 1.1rem;
80
+ text-align: left;
81
+ box-shadow: 0 2px 4px rgba(0, 0, 0, 0.05);
82
  }
83
 
84
  .answer-option:hover {
85
+ background-color: #f0f8ff;
86
+ transform: translateY(-2px);
87
+ border-color: #a6d4fa;
88
  }
89
 
90
  .answer-option.selected {
91
+ background-color: #a6d4fa;
92
+ border-color: #3498db;
93
+ color: #fff;
94
+ box-shadow: 0 4px 6px rgba(0, 0, 0, 0.1);
95
  }
96
 
97
  .answer-option.correct {
98
+ background-color: #86ef7d;
99
+ border-color: #22c55e;
100
  color: #fff;
101
+ font-weight: 600;
102
  }
103
 
104
  .answer-option.incorrect {
105
+ background-color: #fca5a5;
106
+ border-color: #dc2626;
107
  color: #fff;
108
+ font-weight: 600;
109
  }
110
 
111
  #result-message {
112
  font-size: 1.2rem;
113
  margin-bottom: 25px;
114
+ font-weight: 500;
115
  }
116
 
117
  .correct-message {
118
+ color: #22c55e;
119
  }
120
 
121
  .incorrect-message {
122
+ color: #dc2626;
123
  }
124
 
125
+ #next-button, #skip-button, #reset-button, #review-button {
 
126
  padding: 12px 25px;
127
  font-size: 1.1rem;
128
+ margin: 5px 10px;
129
  border-radius: 8px;
130
  border: none;
131
  cursor: pointer;
132
+ transition: background-color 0.3s ease, transform 0.2s ease, box-shadow 0.3s ease;
133
+ font-weight: 500;
134
+ box-shadow: 0 2px 5px rgba(0, 0, 0, 0.1);
135
  }
136
 
137
  #next-button {
138
+ background-color: #3498db;
139
  color: #fff;
140
  }
141
  #next-button:hover {
142
  background-color: #2980b9;
143
  transform: translateY(-2px);
144
+ box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15);
145
  }
146
  #next-button:disabled {
147
+ background-color: #b8c6da;
148
  cursor: not-allowed;
149
+ transform: none;
150
+ box-shadow: none;
151
  }
152
 
153
  #skip-button {
154
+ background-color: #f39c12;
155
  color: #fff;
156
  }
157
  #skip-button:hover {
158
  background-color: #e67e22;
159
  transform: translateY(-2px);
160
+ box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15);
161
  }
162
 
163
  #reset-button {
164
+ background-color: #2ecc71;
165
  color: #fff;
166
  }
167
  #reset-button:hover {
168
  background-color: #27ae60;
169
  transform: translateY(-2px);
170
+ box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15);
171
+ }
172
+
173
+ #review-button {
174
+ background-color: #8e44ad;
175
+ color: #fff;
176
+ }
177
+ #review-button:hover {
178
+ background-color: #732d91;
179
+ transform: translateY(-2px);
180
+ box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15);
181
  }
182
 
183
  #final-score-area {
184
  font-size: 1.5rem;
185
  font-weight: 600;
186
  margin-bottom: 30px;
187
+ color: #2c3e50;
188
  }
189
 
190
  .pass-message {
191
+ color: #22c55e;
192
  }
193
 
194
  .fail-message {
195
+ color: #dc2626;
196
  }
197
 
198
+ #review-section {
199
+ text-align: left;
200
+ margin-top: 20px;
201
+ }
202
+
203
+ .review-item {
204
+ margin-bottom: 20px;
205
+ padding: 15px;
206
+ background-color: #f9f9f9;
207
+ border-radius: 8px;
208
+ border: 1px solid #ddd;
209
+ }
210
+
211
+ .review-question {
212
+ font-size: 1.2rem;
213
+ color: #2c3e50;
214
+ margin-bottom: 10px;
215
+ }
216
+
217
+ .review-answer {
218
+ font-size: 1.1rem;
219
+ margin-bottom: 5px;
220
+ }
221
+
222
+ .review-explanation {
223
+ font-size: 1rem;
224
+ color: #555;
225
+ margin-top: 10px;
226
+ }
227
+
228
+ @media (max-width: 768px) {
229
  #root {
230
  width: 95%;
231
  padding: 20px;
 
239
  font-size: 1.1rem;
240
  padding: 12px;
241
  }
242
+ #next-button, #skip-button, #reset-button, #review-button {
243
  padding: 10px 20px;
244
  font-size: 1rem;
245
  }
 
251
  }
252
  }
253
 
254
+ @media (max-width: 480px) {
255
  #root {
256
  width: 100%;
257
  padding: 15px;
258
+ border-radius: 0;
259
+ box-shadow: none;
260
  }
261
  .answer-option {
262
  padding: 10px;
 
267
  font-size: 1rem;
268
  padding: 10px;
269
  }
270
+ #next-button, #skip-button, #reset-button, #review-button {
271
  padding: 10px 18px;
272
  font-size: 0.9rem;
273
+ margin: 5px 5px;
274
  }
275
  h1 {
276
  font-size: 1.5rem;
 
301
  "To predict stock market trends.",
302
  ],
303
  correctAnswer: "To understand and generate human language.",
304
+ explanation: "Language modeling aims to understand and generate human language by predicting the likelihood of word sequences, forming the basis for many NLP applications."
305
  },
306
  {
307
  question: "Which of the following is NOT a stage in the development of language models?",
 
312
  "Quantum language models (QLM)",
313
  ],
314
  correctAnswer: "Quantum language models (QLM)",
315
+ explanation: "Quantum language models (QLM) are not a recognized stage in language model development; the progression includes SLMs, NLMs, and PLMs."
316
  },
317
  {
318
  question: "What is the basic idea behind statistical language models (SLMs)?",
 
323
  "To pre-train Transformer models.",
324
  ],
325
  correctAnswer: "To build word prediction models based on the Markov assumption.",
326
+ explanation: "SLMs rely on the Markov assumption, which posits that the probability of a word depends only on a fixed number of previous words (n-grams)."
327
  },
328
  {
329
  question: "What is a limitation of SLMs?",
 
334
  "They are not used in NLP.",
335
  ],
336
  correctAnswer: "They suffer from the curse of dimensionality.",
337
+ explanation: "SLMs struggle with high-dimensional data due to the curse of dimensionality, where the number of possible n-grams grows exponentially, making it hard to estimate probabilities accurately."
338
  },
339
  {
340
  question: "What is the main characteristic of neural language models (NLMs)?",
 
345
  "They use n-gram models.",
346
  ],
347
  correctAnswer: "They characterize the probability of word sequences by neural networks.",
348
+ explanation: "NLMs use neural networks to model the probability of word sequences, overcoming some limitations of SLMs by learning continuous representations."
349
  },
350
  {
351
  question: "What concept did the work in [1] introduce?",
 
356
  "Pre-training and fine-tuning.",
357
  ],
358
  correctAnswer: "Distributed representation of words.",
359
+ explanation: "The work in [1] refers to early neural network models (e.g., word2vec) that introduced distributed representations, allowing words to be represented as dense vectors in a continuous space."
360
  },
361
  {
362
  question: "What is word2vec?",
 
367
  "A pre-trained language model.",
368
  ],
369
  correctAnswer: "A simplified shallow neural network for learning distributed word representations.",
370
+ explanation: "Word2vec is a shallow neural network designed to learn distributed word representations (word embeddings) efficiently from large text corpora."
371
  },
372
  {
373
  question: "What was an early attempt at capturing context-aware word representations?",
374
  options: ["BERT", "ELMo", "GPT-2", "word2vec"],
375
  correctAnswer: "ELMo",
376
+ explanation: "ELMo (Embeddings from Language Models) was an early model that captured context-aware word representations by using bidirectional LSTMs, unlike the static embeddings of word2vec."
377
  },
378
  {
379
  question: "Which architecture is highly parallelizable and used in BERT?",
380
  options: ["RNN", "LSTM", "Transformer", "MLP"],
381
  correctAnswer: "Transformer",
382
+ explanation: "BERT uses the Transformer architecture, which is highly parallelizable due to its self-attention mechanism, unlike sequential models like RNNs or LSTMs."
383
  },
384
  {
385
  question: "What is a key feature of BERT?",
 
390
  "It predicts the next word based on the most recent context.",
391
  ],
392
  correctAnswer: "It pre-trains bidirectional language models.",
393
+ explanation: "BERT’s key feature is its bidirectional pre-training, allowing it to consider both left and right context for each word, unlike unidirectional models."
394
  },
395
  {
396
  question: "What paradigm did BERT inspire?",
 
401
  "Distributed representation learning.",
402
  ],
403
  correctAnswer: "Pre-training and fine-tuning.",
404
+ explanation: "BERT popularized the pre-training and fine-tuning paradigm, where a model is pre-trained on a large corpus and then fine-tuned for specific tasks."
405
  },
406
  {
407
  question: "What do researchers find about scaling PLMs?",
 
412
  "It only works for small models.",
413
  ],
414
  correctAnswer: "It often leads to an improved model capacity.",
415
+ explanation: "Scaling pre-trained language models (PLMs) by increasing parameters and data often improves their capacity to handle complex tasks, as seen in models like GPT-3."
416
  },
417
  {
418
  question: "What is a characteristic of large-sized PLMs (LLMs)?",
 
423
  "They are less complex than SLMs.",
424
  ],
425
  correctAnswer: "They display different behaviors from smaller PLMs.",
426
+ explanation: "Large-sized PLMs (LLMs) exhibit emergent behaviors, such as few-shot learning, that smaller PLMs like BERT typically do not show."
427
  },
428
  {
429
  question: "What is an example of an emergent ability in LLMs?",
 
434
  "Using only n-gram models.",
435
  ],
436
  correctAnswer: "Solving few-shot tasks through in-context learning.",
437
+ explanation: "An emergent ability in LLMs is solving few-shot tasks via in-context learning, where the model adapts to new tasks with just a few examples provided in the prompt."
438
  },
439
+ {
440
  question: "What does the term 'LLM' stand for?",
441
  options: [
442
  "Limited Language Model",
 
445
  "Logical Language Model",
446
  ],
447
  correctAnswer: "Large Language Model",
448
+ explanation: "LLM stands for Large Language Model, referring to models with billions of parameters trained on massive datasets."
449
  },
450
  {
451
  question: "Which of the following is a remarkable application of LLMs?",
452
  options: ["ELMo", "ChatGPT", "word2vec", "SLM"],
453
  correctAnswer: "ChatGPT",
454
+ explanation: "ChatGPT, built on the GPT architecture, is a remarkable LLM application known for its conversational abilities and widespread use."
455
  },
456
  {
457
  question: "What is a key focus of the latest language models (e.g., GPT-4)?",
 
462
  "Using only statistical methods.",
463
  ],
464
  correctAnswer: "Complex task solving.",
465
+ explanation: "Latest models like GPT-4 focus on solving complex tasks, leveraging their scale and training to handle diverse, intricate problems."
466
  },
467
  {
468
  question: "How do LLMs differ from small PLMs in accessing them?",
 
473
  "They are accessed using n-gram models.",
474
  ],
475
  correctAnswer: "They are accessed through the prompting interface.",
476
+ explanation: "LLMs are typically accessed via prompting, where users provide instructions or examples in natural language, unlike smaller PLMs that often require fine-tuning."
477
  },
478
  {
479
  question: "What is a challenge in developing LLMs?",
 
484
  "It is very costly to train them due to huge demand for computation resources.",
485
  ],
486
  correctAnswer: "It is very costly to train them due to huge demand for computation resources.",
487
+ explanation: "Training LLMs requires vast computational resources, making it expensive and often limiting development to well-funded organizations."
488
  },
489
  {
490
  question: "What is a potential issue with LLMs despite their capacities?",
 
495
  "They do not require effective control approaches.",
496
  ],
497
  correctAnswer: "They are likely to produce toxic, fictitious, or harmful content.",
498
+ explanation: "Despite their capabilities, LLMs can generate toxic or false content due to biases in training data or lack of perfect alignment with human values."
499
  },
500
  {
501
  question: "What are the four major aspects of LLMs covered in the survey?",
 
506
  "Input, processing, output, and storage.",
507
  ],
508
  correctAnswer: "Pre-training, adaptation, utilization, and capacity evaluation.",
509
+ explanation: "The survey covers pre-training (initial training), adaptation (tuning), utilization (application), and capacity evaluation (performance assessment) as key aspects of LLMs."
510
  },
511
  {
512
  question: "What is the typical parameter size of LLMs?",
 
517
  "Less than 100 parameters.",
518
  ],
519
  correctAnswer: "Billions or hundreds of billions of parameters.",
520
+ explanation: "LLMs typically have billions or hundreds of billions of parameters, enabling their vast capacity, unlike smaller models with millions."
521
  },
522
  {
523
  question: "On what type of data are LLMs typically trained?",
 
528
  "Audio data only.",
529
  ],
530
  correctAnswer: "Massive text data.",
531
+ explanation: "LLMs are trained on massive text corpora, such as web texts, books, and articles, to capture broad language patterns."
532
  },
533
  {
534
  question: "What strong capacity do LLMs exhibit?",
 
539
  "Performing only simple calculations.",
540
  ],
541
  correctAnswer: "Solving complex tasks via text generation.",
542
+ explanation: "LLMs excel at solving complex tasks by generating text, leveraging their understanding of language and context."
543
  },
544
  {
545
  question: "What does the survey introduce as basic background for LLMs?",
 
550
  "Only scaling laws.",
551
  ],
552
  correctAnswer: "Scaling laws, emergent abilities, and key techniques.",
553
+ explanation: "The survey provides background on scaling laws (performance vs. size), emergent abilities (e.g., few-shot learning), and key techniques (e.g., Transformers)."
554
  },
555
  {
556
  question: "What is the key to understanding the development of language models in research history?",
 
561
  "Ignoring the evolution of model capacities.",
562
  ],
563
  correctAnswer: "The leap from language modeling to task solving.",
564
+ explanation: "The shift from merely modeling language (predicting words) to solving tasks (e.g., reasoning, Q&A) marks a pivotal development in language models."
565
  },
566
  {
567
  question: "Which of the following is NOT a characteristic of LLMs?",
 
572
  "Using Transformer language models.",
573
  ],
574
  correctAnswer: "Training on small datasets.",
575
+ explanation: "LLMs are characterized by training on massive datasets, not small ones, which enables their strong performance."
576
  },
577
  {
578
  question: "What is a major difference between LLMs and previous smaller PLMs?",
 
583
  "LLMs do not require large-scale data.",
584
  ],
585
  correctAnswer: "LLMs display surprising emergent abilities.",
586
+ explanation: "LLMs show emergent abilities like in-context learning, which smaller PLMs typically lack due to their scale and training."
587
  },
588
  {
589
  question: "How has the research paradigm shifted towards the use of LLMs in NLP?",
 
594
  "Ignoring pre-training and fine-tuning.",
595
  ],
596
  correctAnswer: "LLMs serve as a general-purpose language task solver.",
597
+ explanation: "The paradigm has shifted to using LLMs as general-purpose solvers for various NLP tasks via prompting, rather than task-specific fine-tuning."
598
  },
599
  {
600
  question: "What is a challenge that traditional search engines face with the advent of LLMs?",
 
605
  "Easier data processing.",
606
  ],
607
  correctAnswer: "New information seeking way through AI chatbots.",
608
+ explanation: "LLM-powered chatbots offer a conversational way to seek information, challenging the keyword-based approach of traditional search engines."
609
  },
610
  {
611
  question: "What is a current trend in computer vision (CV) research related to LLMs?",
 
616
  "Focusing solely on image classification.",
617
  ],
618
  correctAnswer: "Developing ChatGPT-like vision-language models.",
619
+ explanation: "CV research is trending towards multimodal models that combine vision and language, inspired by ChatGPT’s success."
620
  },
621
  {
622
  question: "What is a potential impact of the new wave of LLM technology?",
 
627
  "Slower development of AI algorithms.",
628
  ],
629
  correctAnswer: "A prosperous ecosystem of real-world applications based on LLMs.",
630
+ explanation: "LLMs are fostering a wide range of real-world applications, from chatbots to automation, due to their versatility."
631
  },
632
  {
633
  question: "What is one of the mysterious aspects of LLMs?",
 
638
  "Why they only work on small datasets.",
639
  ],
640
  correctAnswer: "Why emergent abilities occur in LLMs.",
641
+ explanation: "The emergence of abilities like few-shot learning in LLMs is not fully understood, making it a mysterious aspect of their behavior."
642
  },
643
  {
644
  question: "Why is it difficult for the research community to train capable LLMs?",
 
649
  "Because LLMs are mainly trained by academia.",
650
  ],
651
  correctAnswer: "Because it is very costly to carry out repetitive studies.",
652
+ explanation: "Training LLMs requires extensive computational resources, making repetitive studies costly and limiting academic research."
653
  },
654
  {
655
  question: "What is a challenge in aligning LLMs?",
 
660
  "Reducing their computational costs.",
661
  ],
662
  correctAnswer: "Aligning them with human values or preferences.",
663
+ explanation: "Aligning LLMs with human values is challenging due to biases in data and the complexity of defining universal preferences."
664
  },
665
  {
666
  question: "What does the paper 'Planning for AGI and beyond' discuss?",
 
671
  "Only short-term plans for AI development.",
672
  ],
673
  correctAnswer: "Short-term and long-term plans to approach AGI.",
674
+ explanation: "The paper outlines strategies for developing Artificial General Intelligence (AGI), covering both immediate and future steps."
675
  },
676
  {
677
  question: "What is a recent argument about GPT-4?",
 
682
  "It can only perform simple tasks.",
683
  ],
684
  correctAnswer: "It might be considered an early version of an AGI system.",
685
+ explanation: "Some argue GPT-4’s broad capabilities suggest it could be an early AGI, though it lacks full general intelligence."
686
  },
687
  {
688
  question: "How is Microsoft 365 being empowered?",
 
693
  "By using only statistical models.",
694
  ],
695
  correctAnswer: "By LLMs to automate office work.",
696
+ explanation: "Microsoft 365 integrates LLMs (e.g., via Copilot) to automate tasks like writing, summarizing, and data analysis."
697
  },
698
  {
699
+ question: "What is a suggestion for choosing layer normalization in LLMs?",
700
+ options: ["Post RMSNorm", "Pre RMSNorm", "Post LN", "No Normalization"],
701
+ correctAnswer: "Pre RMSNorm",
702
+ explanation: "Pre RMSNorm (Root Mean Square Normalization before layers) is suggested for LLMs due to its stability and performance benefits."
703
  },
704
  {
705
+ question: "Which activation function is recommended for stronger generalization and training stability?",
706
+ options: ["ReLU", "Sigmoid", "SwiGLU", "Tanh"],
707
+ correctAnswer: "SwiGLU",
708
+ explanation: "SwiGLU (Swish-Gated Linear Unit) is recommended for LLMs as it improves generalization and training stability over traditional functions like ReLU."
709
  },
710
  {
711
+ question: "Which position embedding is considered a better choice for LLMs?",
712
+ options: ["Absolute Positional Embedding", "Relative Positional Embedding", "RoPE", "Sinusoidal Positional Encoding"],
713
+ correctAnswer: "RoPE",
714
+ explanation: "RoPE (Rotary Position Embedding) is favored in LLMs for its ability to encode relative positions efficiently and scale with sequence length."
715
  },
716
  {
717
  question: "What is the primary role of pre-training in LLMs?",
718
+ options: [
719
  "To fine-tune models for specific tasks.",
720
  "To encode general knowledge from large-scale corpus.",
721
  "To reduce the size of the model.",
722
  "To improve inference speed.",
723
  ],
724
  correctAnswer: "To encode general knowledge from large-scale corpus.",
725
+ explanation: "Pre-training encodes general knowledge from vast text corpora into LLMs, providing a foundation for later task-specific adaptation."
726
+ },
727
  {
728
  question: "What type of data is included in the arXiv Dataset?",
729
  options: [
 
733
  "Image and video data.",
734
  ],
735
  correctAnswer: "Scientific publication data.",
736
+ explanation: "The arXiv Dataset contains scientific publication data, primarily research papers, used for training models on academic content."
737
  },
738
  {
739
  question: "What is the approximate size of the peS2o dataset?",
740
  options: ["42MB", "42GB", "42TB", "42B tokens"],
741
  correctAnswer: "42B tokens",
742
+ explanation: "The peS2o dataset is approximately 42 billion tokens, a massive text corpus used for training LLMs."
743
  },
744
  {
745
  question: "What is a characteristic of the articles in Wikipedia?",
 
750
  "They cover only a narrow range of topics.",
751
  ],
752
  correctAnswer: "They are composed in an expository style with references.",
753
+ explanation: "Wikipedia articles are written in an expository style, providing detailed explanations with references, making them a valuable training resource."
754
  },
755
  {
756
  question: "What is a technique used to improve memory efficiency and throughput of deployed LLMs?",
757
  options: ["Data Parallelism", "Tensor Parallelism", "Pipeline Parallelism", "PagedAttention"],
758
  correctAnswer: "PagedAttention",
759
+ explanation: "PagedAttention improves memory efficiency and throughput in LLMs by managing key-value caches more effectively during inference."
760
  },
761
  {
762
  question: "How does PagedAttention partition sequences?",
 
767
  "Into overlapping segments.",
768
  ],
769
  correctAnswer: "Into subsequences.",
770
+ explanation: "PagedAttention partitions sequences into subsequences, allowing efficient memory management by processing them in blocks."
771
  },
772
  {
773
  question: "What is the benefit of using PagedAttention?",
 
778
  "Improves model accuracy.",
779
  ],
780
  correctAnswer: "Increases GPU utilization and enables efficient memory sharing.",
781
+ explanation: "PagedAttention boosts GPU utilization and memory sharing, optimizing resource use during LLM inference."
782
  },
783
  {
784
  question: "What type of floating-point number was predominantly used for pre-training in previous PLMs like BERT?",
785
  options: ["FP16", "BF16", "FP32", "INT8"],
786
  correctAnswer: "FP32",
787
+ explanation: "FP32 (32-bit floating-point) was commonly used in earlier PLMs like BERT for its high precision during pre-training."
788
  },
789
  {
790
  question: "Why have some studies started to use FP16 for pre-training LLMs?",
 
795
  "To avoid the loss of computational accuracy.",
796
  ],
797
  correctAnswer: "To reduce memory usage and communication overhead.",
798
+ explanation: "FP16 (16-bit floating-point) reduces memory usage and communication overhead, making pre-training LLMs more efficient despite lower precision."
799
  },
800
  {
801
  question: "What is a potential issue with using FP16 for training?",
 
806
  "Improved model performance.",
807
  ],
808
  correctAnswer: "Loss of computational accuracy.",
809
+ explanation: "FP16’s lower precision can lead to a loss of computational accuracy, potentially affecting model quality during training."
810
  },
811
  {
812
  question: "What is BF16?",
 
817
  "A method for parallel training.",
818
  ],
819
  correctAnswer: "Brain Floating Point, an alternative to FP16.",
820
+ explanation: "BF16 (Brain Floating Point) is a 16-bit format developed by Google, offering a balance between FP16’s efficiency and FP32’s precision."
821
  },
822
  {
823
  question: "How does BF16 compare to FP16 in terms of representation accuracy for pre-training?",
 
828
  "BF16 is not suitable for pre-training.",
829
  ],
830
  correctAnswer: "BF16 generally performs better than FP16.",
831
+ explanation: "BF16 provides better representation accuracy than FP16 due to its wider dynamic range, making it more suitable for pre-training LLMs."
832
  },
833
  {
834
  question: "Which training technique is often used jointly with 3D parallelism to improve training throughput?",
 
839
  "PagedAttention.",
840
  ],
841
  correctAnswer: "Mixed precision training.",
842
+ explanation: "Mixed precision training, combining FP16/BF16 with FP32, is used with 3D parallelism (data, tensor, pipeline) to boost LLM training throughput."
843
  },
844
  {
845
  question: "What type of parallelism was used to train BLOOM on 384 A100 GPUs?",
 
850
  "Only pipeline parallelism.",
851
  ],
852
  correctAnswer: "8-way data parallelism, 4-way tensor parallelism, and 12-way pipeline parallelism.",
853
+ explanation: "BLOOM used a combination of 8-way data, 4-way tensor, and 12-way pipeline parallelism to efficiently train on 384 A100 GPUs."
854
  },
855
  {
856
  question: "What is the primary approach to using LLMs after pre-training or adaptation tuning?",
 
861
  "Using only n-gram models.",
862
  ],
863
  correctAnswer: "Designing suitable prompting strategies.",
864
+ explanation: "Post-pre-training, LLMs are primarily used via prompting strategies, where carefully crafted inputs elicit desired outputs."
865
  },
866
  {
867
  question: "What is a representative prompting method discussed in the text?",
868
  options: ["Fine-tuning", "Backpropagation", "In-context learning", "Random search"],
869
  correctAnswer: "In-context learning",
870
+ explanation: "In-context learning is a key prompting method where LLMs learn tasks from examples provided in the input prompt."
871
  },
872
  {
873
  question: "What does in-context learning involve?",
 
878
  "Employing only manual creation of prompts.",
879
  ],
880
  correctAnswer: "Formulating task description and demonstrations in natural language text.",
881
+ explanation: "In-context learning involves providing a task description and examples in natural language within the prompt to guide the LLM."
882
  },
883
  {
884
  question: "What is the process of manually creating a suitable prompt also called?",
885
  options: ["Automatic prompt optimization", "Prompt engineering", "Prompt tuning", "Prompt generation"],
886
  correctAnswer: "Prompt engineering",
887
+ explanation: "Prompt engineering refers to the manual design of prompts to effectively leverage LLMs’ capabilities for specific tasks."
888
  },
889
  {
890
  question: "What is the impact of a well-designed prompt on LLMs?",
 
895
  "It makes LLMs generate random outputs.",
896
  ],
897
  correctAnswer: "It is very helpful to elicit the abilities of LLMs.",
898
+ explanation: "A well-designed prompt significantly enhances an LLM’s ability to perform tasks by providing clear context and instructions."
899
  },
900
  {
901
  question: "What type of data is considered well-organized with algorithmic logic and programming flow?",
902
  options: ["Natural language text", "Image data", "Code data", "Audio data"],
903
  correctAnswer: "Code data",
904
+ explanation: "Code data is structured with algorithmic logic and programming flow, making it distinct from unstructured natural language text."
905
  },
906
  {
907
  question: "What ability do models trained on code show?",
908
  options: ["Weak reasoning ability", "Strong reasoning ability", "No reasoning ability", "Only language generation ability"],
909
  correctAnswer: "Strong reasoning ability",
910
+ explanation: "Models trained on code exhibit strong reasoning ability due to the logical and structured nature of programming data."
911
  },
912
  {
913
  question: "What is a hypothesis regarding code data and LLMs' reasoning performance?",
 
918
  "Code data is only useful for code generation.",
919
  ],
920
  correctAnswer: "Code data may be useful to improve the reasoning performance of LLMs.",
921
+ explanation: "It’s hypothesized that code data’s logical structure could enhance LLMs’ reasoning skills beyond just code generation."
922
  },
923
  {
924
  question: "What is a characteristic of LLMs' text generation quality?",
 
929
  "It cannot be evaluated.",
930
  ],
931
  correctAnswer: "It is comparable to human-written texts.",
932
+ explanation: "LLMs generate text that is often comparable to human-written content in coherence and quality, a testament to their training scale."
933
  },
934
  {
935
  question: "How can LLMs be used in the context of generation evaluation?",
 
940
  "To perform only statistical analysis.",
941
  ],
942
  correctAnswer: "As language generation evaluators.",
943
+ explanation: "LLMs can evaluate generated text by assessing its quality, coherence, or relevance, supplementing human judgment."
944
  },
945
  {
946
  question: "What is a limitation of LLMs in specialized generation?",
 
951
  "They are only good at generating code.",
952
  ],
953
  correctAnswer: "They have learned general language patterns but underperform in specialized generation.",
954
+ explanation: "LLMs excel in general language but may struggle with highly specialized domains due to limited domain-specific training data."
955
  },
956
  {
957
  question: "What is a common approach to enhancing LLMs' factual knowledge?",
 
962
  "Ignoring up-to-date information.",
963
  ],
964
  correctAnswer: "Incorporating extracted relevant information into the context.",
965
+ explanation: "Enhancing LLMs’ factual knowledge often involves adding relevant external information (e.g., via retrieval-augmented generation) to the context."
966
  },
967
  {
968
  question: "What is a finding about smaller models with instruction tuning compared to larger models without it?",
 
973
  "Larger models are always better.",
974
  ],
975
  correctAnswer: "Smaller models can perform better.",
976
+ explanation: "Smaller models with instruction tuning can outperform larger untuned models by being more aligned with specific tasks."
977
  },
978
  {
979
  question: "What does instruction tuning enable LLMs to do?",
 
984
  "Perform only without demonstrations.",
985
  ],
986
  correctAnswer: "Follow human instructions to perform specific tasks, even on unseen tasks.",
987
+ explanation: "Instruction tuning allows LLMs to generalize to unseen tasks by following human instructions provided in natural language."
988
  },
989
  {
990
  question: "What is a benefit of instruction tuning?",
 
995
  "It does not improve model performance.",
996
  ],
997
  correctAnswer: "It is much less costly than pre-training.",
998
+ explanation: "Instruction tuning is less resource-intensive than pre-training, requiring only a smaller dataset of instructions to adapt the model."
999
  },
1000
  {
1001
  question: "What is the primary goal of instruction tuning?",
 
1006
  "To generate random text",
1007
  ],
1008
  correctAnswer: "To make LLMs better at following instructions",
1009
+ explanation: "The main goal of instruction tuning is to improve LLMs’ ability to accurately follow human instructions for various tasks."
1010
  },
1011
  {
1012
  question: "What kind of tasks does instruction tuning help LLMs perform?",
 
1017
  "No specific tasks",
1018
  ],
1019
  correctAnswer: "Specific tasks without demonstrations",
1020
+ explanation: "Instruction tuning enables LLMs to perform specific tasks based solely on instructions, without needing example demonstrations."
1021
  },
1022
  {
1023
  question: "How does instruction tuning affect LLMs' ability to follow instructions?",
 
1028
  "It makes them ignore instructions",
1029
  ],
1030
  correctAnswer: "It enhances their ability",
1031
+ explanation: "Instruction tuning enhances LLMs’ capability to interpret and act on human instructions effectively."
1032
  },
1033
  {
1034
  question: "What have a large number of studies confirmed about instruction tuning?",
 
1039
  "It decreases performance",
1040
  ],
1041
  correctAnswer: "It achieves superior performance on seen and unseen tasks",
1042
+ explanation: "Studies show instruction tuning boosts LLM performance on both familiar (seen) and new (unseen) tasks."
1043
  },
1044
  {
1045
  question: "What is a key aspect of high-quality long CoT data curation?",
 
1050
  "Manual data collection only",
1051
  ],
1052
  correctAnswer: "Using open models or APIs for data synthesis",
1053
+ explanation: "High-quality Chain-of-Thought (CoT) data is often curated using open models or APIs to synthesize detailed reasoning steps."
1054
  },
1055
  {
1056
  question: "What is the basic idea behind creating long CoT response data?",
 
1061
  "Ignoring the prompt structure",
1062
  ],
1063
  correctAnswer: "Feeding prompts into teacher models",
1064
+ explanation: "Long CoT response data is created by feeding prompts into teacher models to generate step-by-step reasoning responses."
1065
  },
1066
  {
1067
  question: "What is the finding about smaller models with instruction tuning compared to larger models without fine-tuning?",
 
1072
  "Larger models are always better.",
1073
  ],
1074
  correctAnswer: "Smaller models can perform better.",
1075
+ explanation: "Research indicates that smaller, instruction-tuned models can outperform larger models without tuning due to better task alignment."
1076
  },
1077
  ];
1078
 
 
1084
  const [quizEnd, setQuizEnd] = React.useState(false);
1085
  const [selectedAnswerColor, setSelectedAnswerColor] = React.useState('');
1086
  const [currentQuestionNumber, setCurrentQuestionNumber] = React.useState(1);
1087
+ const [incorrectAnswers, setIncorrectAnswers] = React.useState([]);
1088
+ const [showReview, setShowReview] = React.useState(false);
1089
 
1090
  const currentQuestion = quizData[currentQuestionIndex];
1091
  const totalQuestions = quizData.length;
1092
 
 
1093
  const handleAnswerSelection = (answer) => {
1094
  setSelectedAnswer(answer);
1095
  if (answer === currentQuestion.correctAnswer) {
 
1099
  } else {
1100
  setMessage('Incorrect!');
1101
  setSelectedAnswerColor('incorrect');
1102
+ setIncorrectAnswers([...incorrectAnswers, {
1103
+ question: currentQuestion.question,
1104
+ selectedAnswer: answer,
1105
+ correctAnswer: currentQuestion.correctAnswer,
1106
+ explanation: currentQuestion.explanation
1107
+ }]);
1108
  }
1109
  };
1110
 
 
1140
  setQuizEnd(false);
1141
  setSelectedAnswerColor('');
1142
  setCurrentQuestionNumber(1);
1143
+ setIncorrectAnswers([]);
1144
+ setShowReview(false);
1145
+ };
1146
+
1147
+ const handleReviewIncorrect = () => {
1148
+ setShowReview(true);
1149
  };
1150
 
1151
  return (
 
1156
  <>
1157
  <div id="question-area">
1158
  Question {currentQuestionNumber}/{totalQuestions}: {currentQuestion.question}
1159
+ </div>
1160
  <div id="answer-options">
1161
  {currentQuestion.options.map((option) => (
1162
  <div
1163
  key={option}
1164
  className={`answer-option ${selectedAnswer === option ? selectedAnswerColor : ''} ${selectedAnswer ? 'disabled' : ''} ${selectedAnswerColor && option === currentQuestion.correctAnswer ? 'correct' : ''}`}
1165
+ onClick={() => !selectedAnswer && handleAnswerSelection(option)}
1166
  >
1167
  {option}
1168
  </div>
 
1182
  <>
1183
  <div id="final-score-area">
1184
  Your Final Score: {score} / {quizData.length} <br/>
1185
+ {score >= Math.ceil(quizData.length * 0.8) ? (
1186
  <span className="pass-message">Passed</span>
1187
  ) : (
1188
  <span className="fail-message">Failed</span>
1189
  )}
1190
  </div>
1191
+ <div style={{display: 'flex', justifyContent: 'center'}}>
1192
+ <button id="reset-button" onClick={handleResetQuiz}>
1193
+ Play Again
1194
+ </button>
1195
+ {incorrectAnswers.length > 0 && (
1196
+ <button id="review-button" onClick={handleReviewIncorrect}>
1197
+ Review Incorrect Answers
1198
+ </button>
1199
+ )}
1200
+ </div>
1201
+ {showReview && (
1202
+ <div id="review-section">
1203
+ <h2>Review of Incorrect Answers</h2>
1204
+ {incorrectAnswers.map((item, index) => (
1205
+ <div key={index} className="review-item">
1206
+ <div className="review-question">{item.question}</div>
1207
+ <div className="review-answer">
1208
+ <strong>Your Answer:</strong> {item.selectedAnswer} <span className="incorrect-message">(Incorrect)</span>
1209
+ </div>
1210
+ <div className="review-answer">
1211
+ <strong>Correct Answer:</strong> {item.correctAnswer} <span className="correct-message">(Correct)</span>
1212
+ </div>
1213
+ <div className="review-explanation">
1214
+ <strong>Explanation:</strong> {item.explanation}
1215
+ </div>
1216
+ </div>
1217
+ ))}
1218
+ </div>
1219
+ )}
1220
  </>
1221
  )}
1222
  </div>
 
1226
  ReactDOM.render(<QuizApp />, document.getElementById("root"));
1227
  </script>
1228
  </body>
1229
+ </html>