Spaces:
Running
Running
Update index.html
Browse files- index.html +240 -86
index.html
CHANGED
|
@@ -20,173 +20,212 @@
|
|
| 20 |
justify-content: center;
|
| 21 |
align-items: center;
|
| 22 |
min-height: 100vh;
|
| 23 |
-
background-image: url('quiz-background.jpg');
|
| 24 |
background-size: cover;
|
| 25 |
background-position: center;
|
| 26 |
}
|
| 27 |
|
| 28 |
#root {
|
| 29 |
-
background-color: rgba(255, 255, 255, 0.95);
|
| 30 |
padding: 30px;
|
| 31 |
border-radius: 12px;
|
| 32 |
-
box-shadow: 0 8px 20px rgba(0, 0, 0, 0.1);
|
| 33 |
-
width: 80%;
|
| 34 |
-
max-width: 800px;
|
| 35 |
text-align: center;
|
| 36 |
-
box-sizing: border-box;
|
| 37 |
}
|
| 38 |
|
| 39 |
h1 {
|
| 40 |
font-size: 2rem;
|
| 41 |
margin-bottom: 20px;
|
| 42 |
-
color: #3498db;
|
| 43 |
-
font-weight: 600;
|
| 44 |
}
|
| 45 |
|
| 46 |
p {
|
| 47 |
font-size: 1.1rem;
|
| 48 |
-
margin-bottom: 25px;
|
| 49 |
-
color: #555;
|
| 50 |
-
line-height: 1.7;
|
| 51 |
}
|
| 52 |
|
| 53 |
#question-area {
|
| 54 |
font-size: 1.2rem;
|
| 55 |
margin-bottom: 20px;
|
| 56 |
padding: 15px;
|
| 57 |
-
background-color: #e8f0fa;
|
| 58 |
border-radius: 8px;
|
| 59 |
-
border: 1px solid #b8c6da;
|
| 60 |
-
color: #2c3e50;
|
| 61 |
-
text-align: left;
|
| 62 |
}
|
| 63 |
|
| 64 |
#answer-options {
|
| 65 |
display: flex;
|
| 66 |
flex-direction: column;
|
| 67 |
-
align-items: stretch;
|
| 68 |
-
margin-bottom: 25px;
|
| 69 |
}
|
| 70 |
|
| 71 |
.answer-option {
|
| 72 |
padding: 15px;
|
| 73 |
-
margin-bottom: 12px;
|
| 74 |
-
background-color: #fff;
|
| 75 |
border-radius: 8px;
|
| 76 |
-
border: 1px solid #ddd;
|
| 77 |
cursor: pointer;
|
| 78 |
-
transition: background-color 0.3s ease, transform 0.2s ease;
|
| 79 |
-
font-size: 1.1rem;
|
| 80 |
-
text-align: left;
|
| 81 |
-
box-shadow: 0 2px 4px rgba(0, 0, 0, 0.05);
|
| 82 |
}
|
| 83 |
|
| 84 |
.answer-option:hover {
|
| 85 |
-
background-color: #f0f8ff;
|
| 86 |
-
transform: translateY(-2px);
|
| 87 |
-
border-color: #a6d4fa;
|
| 88 |
}
|
| 89 |
|
| 90 |
.answer-option.selected {
|
| 91 |
-
background-color: #a6d4fa;
|
| 92 |
-
border-color: #3498db;
|
| 93 |
-
color: #fff;
|
| 94 |
-
box-shadow: 0 4px 6px rgba(0, 0, 0, 0.1);
|
| 95 |
}
|
| 96 |
|
| 97 |
.answer-option.correct {
|
| 98 |
-
background-color: #86ef7d;
|
| 99 |
-
border-color: #22c55e;
|
| 100 |
color: #fff;
|
| 101 |
-
font-weight: 600;
|
| 102 |
}
|
| 103 |
|
| 104 |
.answer-option.incorrect {
|
| 105 |
-
background-color: #fca5a5;
|
| 106 |
-
border-color: #dc2626;
|
| 107 |
color: #fff;
|
| 108 |
-
font-weight: 600;
|
| 109 |
}
|
| 110 |
|
| 111 |
#result-message {
|
| 112 |
font-size: 1.2rem;
|
| 113 |
margin-bottom: 25px;
|
| 114 |
-
font-weight: 500;
|
| 115 |
}
|
| 116 |
|
| 117 |
.correct-message {
|
| 118 |
-
color: #22c55e;
|
| 119 |
}
|
| 120 |
|
| 121 |
.incorrect-message {
|
| 122 |
-
color: #dc2626;
|
| 123 |
}
|
| 124 |
|
| 125 |
-
|
| 126 |
-
#next-button, #skip-button, #reset-button {
|
| 127 |
padding: 12px 25px;
|
| 128 |
font-size: 1.1rem;
|
| 129 |
-
margin: 5px 10px;
|
| 130 |
border-radius: 8px;
|
| 131 |
border: none;
|
| 132 |
cursor: pointer;
|
| 133 |
-
transition: background-color 0.3s ease, transform 0.2s ease, box-shadow 0.3s ease;
|
| 134 |
-
font-weight: 500;
|
| 135 |
-
box-shadow: 0 2px 5px rgba(0, 0, 0, 0.1);
|
| 136 |
}
|
| 137 |
|
| 138 |
#next-button {
|
| 139 |
-
background-color: #3498db;
|
| 140 |
color: #fff;
|
| 141 |
}
|
| 142 |
#next-button:hover {
|
| 143 |
background-color: #2980b9;
|
| 144 |
transform: translateY(-2px);
|
| 145 |
-
box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15);
|
| 146 |
}
|
| 147 |
#next-button:disabled {
|
| 148 |
-
background-color: #b8c6da;
|
| 149 |
cursor: not-allowed;
|
| 150 |
-
transform: none;
|
| 151 |
-
box-shadow: none;
|
| 152 |
}
|
| 153 |
|
| 154 |
#skip-button {
|
| 155 |
-
background-color: #f39c12;
|
| 156 |
color: #fff;
|
| 157 |
}
|
| 158 |
#skip-button:hover {
|
| 159 |
background-color: #e67e22;
|
| 160 |
transform: translateY(-2px);
|
| 161 |
-
box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15);
|
| 162 |
}
|
| 163 |
|
| 164 |
#reset-button {
|
| 165 |
-
background-color: #2ecc71;
|
| 166 |
color: #fff;
|
| 167 |
}
|
| 168 |
#reset-button:hover {
|
| 169 |
background-color: #27ae60;
|
| 170 |
transform: translateY(-2px);
|
| 171 |
-
box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15);
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 172 |
}
|
| 173 |
|
| 174 |
#final-score-area {
|
| 175 |
font-size: 1.5rem;
|
| 176 |
font-weight: 600;
|
| 177 |
margin-bottom: 30px;
|
| 178 |
-
color: #2c3e50;
|
| 179 |
}
|
| 180 |
|
| 181 |
.pass-message {
|
| 182 |
-
color: #22c55e;
|
| 183 |
}
|
| 184 |
|
| 185 |
.fail-message {
|
| 186 |
-
color: #dc2626;
|
| 187 |
}
|
| 188 |
|
| 189 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 190 |
#root {
|
| 191 |
width: 95%;
|
| 192 |
padding: 20px;
|
|
@@ -200,7 +239,7 @@
|
|
| 200 |
font-size: 1.1rem;
|
| 201 |
padding: 12px;
|
| 202 |
}
|
| 203 |
-
#next-button, #skip-button, #reset-button {
|
| 204 |
padding: 10px 20px;
|
| 205 |
font-size: 1rem;
|
| 206 |
}
|
|
@@ -212,12 +251,12 @@
|
|
| 212 |
}
|
| 213 |
}
|
| 214 |
|
| 215 |
-
@media (max-width: 480px) {
|
| 216 |
#root {
|
| 217 |
width: 100%;
|
| 218 |
padding: 15px;
|
| 219 |
-
border-radius: 0;
|
| 220 |
-
box-shadow: none;
|
| 221 |
}
|
| 222 |
.answer-option {
|
| 223 |
padding: 10px;
|
|
@@ -228,10 +267,10 @@
|
|
| 228 |
font-size: 1rem;
|
| 229 |
padding: 10px;
|
| 230 |
}
|
| 231 |
-
#next-button, #skip-button, #reset-button {
|
| 232 |
padding: 10px 18px;
|
| 233 |
font-size: 0.9rem;
|
| 234 |
-
margin: 5px 5px;
|
| 235 |
}
|
| 236 |
h1 {
|
| 237 |
font-size: 1.5rem;
|
|
@@ -262,6 +301,7 @@
|
|
| 262 |
"To predict stock market trends.",
|
| 263 |
],
|
| 264 |
correctAnswer: "To understand and generate human language.",
|
|
|
|
| 265 |
},
|
| 266 |
{
|
| 267 |
question: "Which of the following is NOT a stage in the development of language models?",
|
|
@@ -272,6 +312,7 @@
|
|
| 272 |
"Quantum language models (QLM)",
|
| 273 |
],
|
| 274 |
correctAnswer: "Quantum language models (QLM)",
|
|
|
|
| 275 |
},
|
| 276 |
{
|
| 277 |
question: "What is the basic idea behind statistical language models (SLMs)?",
|
|
@@ -282,6 +323,7 @@
|
|
| 282 |
"To pre-train Transformer models.",
|
| 283 |
],
|
| 284 |
correctAnswer: "To build word prediction models based on the Markov assumption.",
|
|
|
|
| 285 |
},
|
| 286 |
{
|
| 287 |
question: "What is a limitation of SLMs?",
|
|
@@ -292,6 +334,7 @@
|
|
| 292 |
"They are not used in NLP.",
|
| 293 |
],
|
| 294 |
correctAnswer: "They suffer from the curse of dimensionality.",
|
|
|
|
| 295 |
},
|
| 296 |
{
|
| 297 |
question: "What is the main characteristic of neural language models (NLMs)?",
|
|
@@ -302,6 +345,7 @@
|
|
| 302 |
"They use n-gram models.",
|
| 303 |
],
|
| 304 |
correctAnswer: "They characterize the probability of word sequences by neural networks.",
|
|
|
|
| 305 |
},
|
| 306 |
{
|
| 307 |
question: "What concept did the work in [1] introduce?",
|
|
@@ -312,6 +356,7 @@
|
|
| 312 |
"Pre-training and fine-tuning.",
|
| 313 |
],
|
| 314 |
correctAnswer: "Distributed representation of words.",
|
|
|
|
| 315 |
},
|
| 316 |
{
|
| 317 |
question: "What is word2vec?",
|
|
@@ -322,16 +367,19 @@
|
|
| 322 |
"A pre-trained language model.",
|
| 323 |
],
|
| 324 |
correctAnswer: "A simplified shallow neural network for learning distributed word representations.",
|
|
|
|
| 325 |
},
|
| 326 |
{
|
| 327 |
question: "What was an early attempt at capturing context-aware word representations?",
|
| 328 |
options: ["BERT", "ELMo", "GPT-2", "word2vec"],
|
| 329 |
correctAnswer: "ELMo",
|
|
|
|
| 330 |
},
|
| 331 |
{
|
| 332 |
question: "Which architecture is highly parallelizable and used in BERT?",
|
| 333 |
options: ["RNN", "LSTM", "Transformer", "MLP"],
|
| 334 |
correctAnswer: "Transformer",
|
|
|
|
| 335 |
},
|
| 336 |
{
|
| 337 |
question: "What is a key feature of BERT?",
|
|
@@ -342,6 +390,7 @@
|
|
| 342 |
"It predicts the next word based on the most recent context.",
|
| 343 |
],
|
| 344 |
correctAnswer: "It pre-trains bidirectional language models.",
|
|
|
|
| 345 |
},
|
| 346 |
{
|
| 347 |
question: "What paradigm did BERT inspire?",
|
|
@@ -352,6 +401,7 @@
|
|
| 352 |
"Distributed representation learning.",
|
| 353 |
],
|
| 354 |
correctAnswer: "Pre-training and fine-tuning.",
|
|
|
|
| 355 |
},
|
| 356 |
{
|
| 357 |
question: "What do researchers find about scaling PLMs?",
|
|
@@ -362,6 +412,7 @@
|
|
| 362 |
"It only works for small models.",
|
| 363 |
],
|
| 364 |
correctAnswer: "It often leads to an improved model capacity.",
|
|
|
|
| 365 |
},
|
| 366 |
{
|
| 367 |
question: "What is a characteristic of large-sized PLMs (LLMs)?",
|
|
@@ -372,6 +423,7 @@
|
|
| 372 |
"They are less complex than SLMs.",
|
| 373 |
],
|
| 374 |
correctAnswer: "They display different behaviors from smaller PLMs.",
|
|
|
|
| 375 |
},
|
| 376 |
{
|
| 377 |
question: "What is an example of an emergent ability in LLMs?",
|
|
@@ -382,8 +434,9 @@
|
|
| 382 |
"Using only n-gram models.",
|
| 383 |
],
|
| 384 |
correctAnswer: "Solving few-shot tasks through in-context learning.",
|
|
|
|
| 385 |
},
|
| 386 |
-
|
| 387 |
question: "What does the term 'LLM' stand for?",
|
| 388 |
options: [
|
| 389 |
"Limited Language Model",
|
|
@@ -392,11 +445,13 @@
|
|
| 392 |
"Logical Language Model",
|
| 393 |
],
|
| 394 |
correctAnswer: "Large Language Model",
|
|
|
|
| 395 |
},
|
| 396 |
{
|
| 397 |
question: "Which of the following is a remarkable application of LLMs?",
|
| 398 |
options: ["ELMo", "ChatGPT", "word2vec", "SLM"],
|
| 399 |
correctAnswer: "ChatGPT",
|
|
|
|
| 400 |
},
|
| 401 |
{
|
| 402 |
question: "What is a key focus of the latest language models (e.g., GPT-4)?",
|
|
@@ -407,6 +462,7 @@
|
|
| 407 |
"Using only statistical methods.",
|
| 408 |
],
|
| 409 |
correctAnswer: "Complex task solving.",
|
|
|
|
| 410 |
},
|
| 411 |
{
|
| 412 |
question: "How do LLMs differ from small PLMs in accessing them?",
|
|
@@ -417,6 +473,7 @@
|
|
| 417 |
"They are accessed using n-gram models.",
|
| 418 |
],
|
| 419 |
correctAnswer: "They are accessed through the prompting interface.",
|
|
|
|
| 420 |
},
|
| 421 |
{
|
| 422 |
question: "What is a challenge in developing LLMs?",
|
|
@@ -427,6 +484,7 @@
|
|
| 427 |
"It is very costly to train them due to huge demand for computation resources.",
|
| 428 |
],
|
| 429 |
correctAnswer: "It is very costly to train them due to huge demand for computation resources.",
|
|
|
|
| 430 |
},
|
| 431 |
{
|
| 432 |
question: "What is a potential issue with LLMs despite their capacities?",
|
|
@@ -437,6 +495,7 @@
|
|
| 437 |
"They do not require effective control approaches.",
|
| 438 |
],
|
| 439 |
correctAnswer: "They are likely to produce toxic, fictitious, or harmful content.",
|
|
|
|
| 440 |
},
|
| 441 |
{
|
| 442 |
question: "What are the four major aspects of LLMs covered in the survey?",
|
|
@@ -447,6 +506,7 @@
|
|
| 447 |
"Input, processing, output, and storage.",
|
| 448 |
],
|
| 449 |
correctAnswer: "Pre-training, adaptation, utilization, and capacity evaluation.",
|
|
|
|
| 450 |
},
|
| 451 |
{
|
| 452 |
question: "What is the typical parameter size of LLMs?",
|
|
@@ -457,6 +517,7 @@
|
|
| 457 |
"Less than 100 parameters.",
|
| 458 |
],
|
| 459 |
correctAnswer: "Billions or hundreds of billions of parameters.",
|
|
|
|
| 460 |
},
|
| 461 |
{
|
| 462 |
question: "On what type of data are LLMs typically trained?",
|
|
@@ -467,6 +528,7 @@
|
|
| 467 |
"Audio data only.",
|
| 468 |
],
|
| 469 |
correctAnswer: "Massive text data.",
|
|
|
|
| 470 |
},
|
| 471 |
{
|
| 472 |
question: "What strong capacity do LLMs exhibit?",
|
|
@@ -477,6 +539,7 @@
|
|
| 477 |
"Performing only simple calculations.",
|
| 478 |
],
|
| 479 |
correctAnswer: "Solving complex tasks via text generation.",
|
|
|
|
| 480 |
},
|
| 481 |
{
|
| 482 |
question: "What does the survey introduce as basic background for LLMs?",
|
|
@@ -487,6 +550,7 @@
|
|
| 487 |
"Only scaling laws.",
|
| 488 |
],
|
| 489 |
correctAnswer: "Scaling laws, emergent abilities, and key techniques.",
|
|
|
|
| 490 |
},
|
| 491 |
{
|
| 492 |
question: "What is the key to understanding the development of language models in research history?",
|
|
@@ -497,6 +561,7 @@
|
|
| 497 |
"Ignoring the evolution of model capacities.",
|
| 498 |
],
|
| 499 |
correctAnswer: "The leap from language modeling to task solving.",
|
|
|
|
| 500 |
},
|
| 501 |
{
|
| 502 |
question: "Which of the following is NOT a characteristic of LLMs?",
|
|
@@ -507,6 +572,7 @@
|
|
| 507 |
"Using Transformer language models.",
|
| 508 |
],
|
| 509 |
correctAnswer: "Training on small datasets.",
|
|
|
|
| 510 |
},
|
| 511 |
{
|
| 512 |
question: "What is a major difference between LLMs and previous smaller PLMs?",
|
|
@@ -517,6 +583,7 @@
|
|
| 517 |
"LLMs do not require large-scale data.",
|
| 518 |
],
|
| 519 |
correctAnswer: "LLMs display surprising emergent abilities.",
|
|
|
|
| 520 |
},
|
| 521 |
{
|
| 522 |
question: "How has the research paradigm shifted towards the use of LLMs in NLP?",
|
|
@@ -527,6 +594,7 @@
|
|
| 527 |
"Ignoring pre-training and fine-tuning.",
|
| 528 |
],
|
| 529 |
correctAnswer: "LLMs serve as a general-purpose language task solver.",
|
|
|
|
| 530 |
},
|
| 531 |
{
|
| 532 |
question: "What is a challenge that traditional search engines face with the advent of LLMs?",
|
|
@@ -537,6 +605,7 @@
|
|
| 537 |
"Easier data processing.",
|
| 538 |
],
|
| 539 |
correctAnswer: "New information seeking way through AI chatbots.",
|
|
|
|
| 540 |
},
|
| 541 |
{
|
| 542 |
question: "What is a current trend in computer vision (CV) research related to LLMs?",
|
|
@@ -547,6 +616,7 @@
|
|
| 547 |
"Focusing solely on image classification.",
|
| 548 |
],
|
| 549 |
correctAnswer: "Developing ChatGPT-like vision-language models.",
|
|
|
|
| 550 |
},
|
| 551 |
{
|
| 552 |
question: "What is a potential impact of the new wave of LLM technology?",
|
|
@@ -557,6 +627,7 @@
|
|
| 557 |
"Slower development of AI algorithms.",
|
| 558 |
],
|
| 559 |
correctAnswer: "A prosperous ecosystem of real-world applications based on LLMs.",
|
|
|
|
| 560 |
},
|
| 561 |
{
|
| 562 |
question: "What is one of the mysterious aspects of LLMs?",
|
|
@@ -567,6 +638,7 @@
|
|
| 567 |
"Why they only work on small datasets.",
|
| 568 |
],
|
| 569 |
correctAnswer: "Why emergent abilities occur in LLMs.",
|
|
|
|
| 570 |
},
|
| 571 |
{
|
| 572 |
question: "Why is it difficult for the research community to train capable LLMs?",
|
|
@@ -577,6 +649,7 @@
|
|
| 577 |
"Because LLMs are mainly trained by academia.",
|
| 578 |
],
|
| 579 |
correctAnswer: "Because it is very costly to carry out repetitive studies.",
|
|
|
|
| 580 |
},
|
| 581 |
{
|
| 582 |
question: "What is a challenge in aligning LLMs?",
|
|
@@ -587,6 +660,7 @@
|
|
| 587 |
"Reducing their computational costs.",
|
| 588 |
],
|
| 589 |
correctAnswer: "Aligning them with human values or preferences.",
|
|
|
|
| 590 |
},
|
| 591 |
{
|
| 592 |
question: "What does the paper 'Planning for AGI and beyond' discuss?",
|
|
@@ -597,6 +671,7 @@
|
|
| 597 |
"Only short-term plans for AI development.",
|
| 598 |
],
|
| 599 |
correctAnswer: "Short-term and long-term plans to approach AGI.",
|
|
|
|
| 600 |
},
|
| 601 |
{
|
| 602 |
question: "What is a recent argument about GPT-4?",
|
|
@@ -607,6 +682,7 @@
|
|
| 607 |
"It can only perform simple tasks.",
|
| 608 |
],
|
| 609 |
correctAnswer: "It might be considered an early version of an AGI system.",
|
|
|
|
| 610 |
},
|
| 611 |
{
|
| 612 |
question: "How is Microsoft 365 being empowered?",
|
|
@@ -617,32 +693,37 @@
|
|
| 617 |
"By using only statistical models.",
|
| 618 |
],
|
| 619 |
correctAnswer: "By LLMs to automate office work.",
|
|
|
|
| 620 |
},
|
| 621 |
{
|
| 622 |
-
|
| 623 |
-
|
| 624 |
-
|
|
|
|
| 625 |
},
|
| 626 |
{
|
| 627 |
-
|
| 628 |
-
|
| 629 |
-
|
|
|
|
| 630 |
},
|
| 631 |
{
|
| 632 |
-
|
| 633 |
-
|
| 634 |
-
|
|
|
|
| 635 |
},
|
| 636 |
{
|
| 637 |
question: "What is the primary role of pre-training in LLMs?",
|
| 638 |
-
options: [
|
| 639 |
"To fine-tune models for specific tasks.",
|
| 640 |
"To encode general knowledge from large-scale corpus.",
|
| 641 |
"To reduce the size of the model.",
|
| 642 |
"To improve inference speed.",
|
| 643 |
],
|
| 644 |
correctAnswer: "To encode general knowledge from large-scale corpus.",
|
| 645 |
-
|
|
|
|
| 646 |
{
|
| 647 |
question: "What type of data is included in the arXiv Dataset?",
|
| 648 |
options: [
|
|
@@ -652,11 +733,13 @@ options: [
|
|
| 652 |
"Image and video data.",
|
| 653 |
],
|
| 654 |
correctAnswer: "Scientific publication data.",
|
|
|
|
| 655 |
},
|
| 656 |
{
|
| 657 |
question: "What is the approximate size of the peS2o dataset?",
|
| 658 |
options: ["42MB", "42GB", "42TB", "42B tokens"],
|
| 659 |
correctAnswer: "42B tokens",
|
|
|
|
| 660 |
},
|
| 661 |
{
|
| 662 |
question: "What is a characteristic of the articles in Wikipedia?",
|
|
@@ -667,11 +750,13 @@ options: [
|
|
| 667 |
"They cover only a narrow range of topics.",
|
| 668 |
],
|
| 669 |
correctAnswer: "They are composed in an expository style with references.",
|
|
|
|
| 670 |
},
|
| 671 |
{
|
| 672 |
question: "What is a technique used to improve memory efficiency and throughput of deployed LLMs?",
|
| 673 |
options: ["Data Parallelism", "Tensor Parallelism", "Pipeline Parallelism", "PagedAttention"],
|
| 674 |
correctAnswer: "PagedAttention",
|
|
|
|
| 675 |
},
|
| 676 |
{
|
| 677 |
question: "How does PagedAttention partition sequences?",
|
|
@@ -682,6 +767,7 @@ options: [
|
|
| 682 |
"Into overlapping segments.",
|
| 683 |
],
|
| 684 |
correctAnswer: "Into subsequences.",
|
|
|
|
| 685 |
},
|
| 686 |
{
|
| 687 |
question: "What is the benefit of using PagedAttention?",
|
|
@@ -692,11 +778,13 @@ options: [
|
|
| 692 |
"Improves model accuracy.",
|
| 693 |
],
|
| 694 |
correctAnswer: "Increases GPU utilization and enables efficient memory sharing.",
|
|
|
|
| 695 |
},
|
| 696 |
{
|
| 697 |
question: "What type of floating-point number was predominantly used for pre-training in previous PLMs like BERT?",
|
| 698 |
options: ["FP16", "BF16", "FP32", "INT8"],
|
| 699 |
correctAnswer: "FP32",
|
|
|
|
| 700 |
},
|
| 701 |
{
|
| 702 |
question: "Why have some studies started to use FP16 for pre-training LLMs?",
|
|
@@ -707,6 +795,7 @@ options: [
|
|
| 707 |
"To avoid the loss of computational accuracy.",
|
| 708 |
],
|
| 709 |
correctAnswer: "To reduce memory usage and communication overhead.",
|
|
|
|
| 710 |
},
|
| 711 |
{
|
| 712 |
question: "What is a potential issue with using FP16 for training?",
|
|
@@ -717,6 +806,7 @@ options: [
|
|
| 717 |
"Improved model performance.",
|
| 718 |
],
|
| 719 |
correctAnswer: "Loss of computational accuracy.",
|
|
|
|
| 720 |
},
|
| 721 |
{
|
| 722 |
question: "What is BF16?",
|
|
@@ -727,6 +817,7 @@ options: [
|
|
| 727 |
"A method for parallel training.",
|
| 728 |
],
|
| 729 |
correctAnswer: "Brain Floating Point, an alternative to FP16.",
|
|
|
|
| 730 |
},
|
| 731 |
{
|
| 732 |
question: "How does BF16 compare to FP16 in terms of representation accuracy for pre-training?",
|
|
@@ -737,6 +828,7 @@ options: [
|
|
| 737 |
"BF16 is not suitable for pre-training.",
|
| 738 |
],
|
| 739 |
correctAnswer: "BF16 generally performs better than FP16.",
|
|
|
|
| 740 |
},
|
| 741 |
{
|
| 742 |
question: "Which training technique is often used jointly with 3D parallelism to improve training throughput?",
|
|
@@ -747,6 +839,7 @@ options: [
|
|
| 747 |
"PagedAttention.",
|
| 748 |
],
|
| 749 |
correctAnswer: "Mixed precision training.",
|
|
|
|
| 750 |
},
|
| 751 |
{
|
| 752 |
question: "What type of parallelism was used to train BLOOM on 384 A100 GPUs?",
|
|
@@ -757,6 +850,7 @@ options: [
|
|
| 757 |
"Only pipeline parallelism.",
|
| 758 |
],
|
| 759 |
correctAnswer: "8-way data parallelism, 4-way tensor parallelism, and 12-way pipeline parallelism.",
|
|
|
|
| 760 |
},
|
| 761 |
{
|
| 762 |
question: "What is the primary approach to using LLMs after pre-training or adaptation tuning?",
|
|
@@ -767,11 +861,13 @@ options: [
|
|
| 767 |
"Using only n-gram models.",
|
| 768 |
],
|
| 769 |
correctAnswer: "Designing suitable prompting strategies.",
|
|
|
|
| 770 |
},
|
| 771 |
{
|
| 772 |
question: "What is a representative prompting method discussed in the text?",
|
| 773 |
options: ["Fine-tuning", "Backpropagation", "In-context learning", "Random search"],
|
| 774 |
correctAnswer: "In-context learning",
|
|
|
|
| 775 |
},
|
| 776 |
{
|
| 777 |
question: "What does in-context learning involve?",
|
|
@@ -782,11 +878,13 @@ options: [
|
|
| 782 |
"Employing only manual creation of prompts.",
|
| 783 |
],
|
| 784 |
correctAnswer: "Formulating task description and demonstrations in natural language text.",
|
|
|
|
| 785 |
},
|
| 786 |
{
|
| 787 |
question: "What is the process of manually creating a suitable prompt also called?",
|
| 788 |
options: ["Automatic prompt optimization", "Prompt engineering", "Prompt tuning", "Prompt generation"],
|
| 789 |
correctAnswer: "Prompt engineering",
|
|
|
|
| 790 |
},
|
| 791 |
{
|
| 792 |
question: "What is the impact of a well-designed prompt on LLMs?",
|
|
@@ -797,16 +895,19 @@ options: [
|
|
| 797 |
"It makes LLMs generate random outputs.",
|
| 798 |
],
|
| 799 |
correctAnswer: "It is very helpful to elicit the abilities of LLMs.",
|
|
|
|
| 800 |
},
|
| 801 |
{
|
| 802 |
question: "What type of data is considered well-organized with algorithmic logic and programming flow?",
|
| 803 |
options: ["Natural language text", "Image data", "Code data", "Audio data"],
|
| 804 |
correctAnswer: "Code data",
|
|
|
|
| 805 |
},
|
| 806 |
{
|
| 807 |
question: "What ability do models trained on code show?",
|
| 808 |
options: ["Weak reasoning ability", "Strong reasoning ability", "No reasoning ability", "Only language generation ability"],
|
| 809 |
correctAnswer: "Strong reasoning ability",
|
|
|
|
| 810 |
},
|
| 811 |
{
|
| 812 |
question: "What is a hypothesis regarding code data and LLMs' reasoning performance?",
|
|
@@ -817,6 +918,7 @@ options: [
|
|
| 817 |
"Code data is only useful for code generation.",
|
| 818 |
],
|
| 819 |
correctAnswer: "Code data may be useful to improve the reasoning performance of LLMs.",
|
|
|
|
| 820 |
},
|
| 821 |
{
|
| 822 |
question: "What is a characteristic of LLMs' text generation quality?",
|
|
@@ -827,6 +929,7 @@ options: [
|
|
| 827 |
"It cannot be evaluated.",
|
| 828 |
],
|
| 829 |
correctAnswer: "It is comparable to human-written texts.",
|
|
|
|
| 830 |
},
|
| 831 |
{
|
| 832 |
question: "How can LLMs be used in the context of generation evaluation?",
|
|
@@ -837,6 +940,7 @@ options: [
|
|
| 837 |
"To perform only statistical analysis.",
|
| 838 |
],
|
| 839 |
correctAnswer: "As language generation evaluators.",
|
|
|
|
| 840 |
},
|
| 841 |
{
|
| 842 |
question: "What is a limitation of LLMs in specialized generation?",
|
|
@@ -847,6 +951,7 @@ options: [
|
|
| 847 |
"They are only good at generating code.",
|
| 848 |
],
|
| 849 |
correctAnswer: "They have learned general language patterns but underperform in specialized generation.",
|
|
|
|
| 850 |
},
|
| 851 |
{
|
| 852 |
question: "What is a common approach to enhancing LLMs' factual knowledge?",
|
|
@@ -857,6 +962,7 @@ options: [
|
|
| 857 |
"Ignoring up-to-date information.",
|
| 858 |
],
|
| 859 |
correctAnswer: "Incorporating extracted relevant information into the context.",
|
|
|
|
| 860 |
},
|
| 861 |
{
|
| 862 |
question: "What is a finding about smaller models with instruction tuning compared to larger models without it?",
|
|
@@ -867,6 +973,7 @@ options: [
|
|
| 867 |
"Larger models are always better.",
|
| 868 |
],
|
| 869 |
correctAnswer: "Smaller models can perform better.",
|
|
|
|
| 870 |
},
|
| 871 |
{
|
| 872 |
question: "What does instruction tuning enable LLMs to do?",
|
|
@@ -877,6 +984,7 @@ options: [
|
|
| 877 |
"Perform only without demonstrations.",
|
| 878 |
],
|
| 879 |
correctAnswer: "Follow human instructions to perform specific tasks, even on unseen tasks.",
|
|
|
|
| 880 |
},
|
| 881 |
{
|
| 882 |
question: "What is a benefit of instruction tuning?",
|
|
@@ -887,6 +995,7 @@ options: [
|
|
| 887 |
"It does not improve model performance.",
|
| 888 |
],
|
| 889 |
correctAnswer: "It is much less costly than pre-training.",
|
|
|
|
| 890 |
},
|
| 891 |
{
|
| 892 |
question: "What is the primary goal of instruction tuning?",
|
|
@@ -897,6 +1006,7 @@ options: [
|
|
| 897 |
"To generate random text",
|
| 898 |
],
|
| 899 |
correctAnswer: "To make LLMs better at following instructions",
|
|
|
|
| 900 |
},
|
| 901 |
{
|
| 902 |
question: "What kind of tasks does instruction tuning help LLMs perform?",
|
|
@@ -907,6 +1017,7 @@ options: [
|
|
| 907 |
"No specific tasks",
|
| 908 |
],
|
| 909 |
correctAnswer: "Specific tasks without demonstrations",
|
|
|
|
| 910 |
},
|
| 911 |
{
|
| 912 |
question: "How does instruction tuning affect LLMs' ability to follow instructions?",
|
|
@@ -917,6 +1028,7 @@ options: [
|
|
| 917 |
"It makes them ignore instructions",
|
| 918 |
],
|
| 919 |
correctAnswer: "It enhances their ability",
|
|
|
|
| 920 |
},
|
| 921 |
{
|
| 922 |
question: "What have a large number of studies confirmed about instruction tuning?",
|
|
@@ -927,6 +1039,7 @@ options: [
|
|
| 927 |
"It decreases performance",
|
| 928 |
],
|
| 929 |
correctAnswer: "It achieves superior performance on seen and unseen tasks",
|
|
|
|
| 930 |
},
|
| 931 |
{
|
| 932 |
question: "What is a key aspect of high-quality long CoT data curation?",
|
|
@@ -937,6 +1050,7 @@ options: [
|
|
| 937 |
"Manual data collection only",
|
| 938 |
],
|
| 939 |
correctAnswer: "Using open models or APIs for data synthesis",
|
|
|
|
| 940 |
},
|
| 941 |
{
|
| 942 |
question: "What is the basic idea behind creating long CoT response data?",
|
|
@@ -947,6 +1061,7 @@ options: [
|
|
| 947 |
"Ignoring the prompt structure",
|
| 948 |
],
|
| 949 |
correctAnswer: "Feeding prompts into teacher models",
|
|
|
|
| 950 |
},
|
| 951 |
{
|
| 952 |
question: "What is the finding about smaller models with instruction tuning compared to larger models without fine-tuning?",
|
|
@@ -957,6 +1072,7 @@ options: [
|
|
| 957 |
"Larger models are always better.",
|
| 958 |
],
|
| 959 |
correctAnswer: "Smaller models can perform better.",
|
|
|
|
| 960 |
},
|
| 961 |
];
|
| 962 |
|
|
@@ -968,12 +1084,12 @@ options: [
|
|
| 968 |
const [quizEnd, setQuizEnd] = React.useState(false);
|
| 969 |
const [selectedAnswerColor, setSelectedAnswerColor] = React.useState('');
|
| 970 |
const [currentQuestionNumber, setCurrentQuestionNumber] = React.useState(1);
|
| 971 |
-
|
|
|
|
| 972 |
|
| 973 |
const currentQuestion = quizData[currentQuestionIndex];
|
| 974 |
const totalQuestions = quizData.length;
|
| 975 |
|
| 976 |
-
|
| 977 |
const handleAnswerSelection = (answer) => {
|
| 978 |
setSelectedAnswer(answer);
|
| 979 |
if (answer === currentQuestion.correctAnswer) {
|
|
@@ -983,6 +1099,12 @@ options: [
|
|
| 983 |
} else {
|
| 984 |
setMessage('Incorrect!');
|
| 985 |
setSelectedAnswerColor('incorrect');
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 986 |
}
|
| 987 |
};
|
| 988 |
|
|
@@ -1018,6 +1140,12 @@ options: [
|
|
| 1018 |
setQuizEnd(false);
|
| 1019 |
setSelectedAnswerColor('');
|
| 1020 |
setCurrentQuestionNumber(1);
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1021 |
};
|
| 1022 |
|
| 1023 |
return (
|
|
@@ -1028,13 +1156,13 @@ options: [
|
|
| 1028 |
<>
|
| 1029 |
<div id="question-area">
|
| 1030 |
Question {currentQuestionNumber}/{totalQuestions}: {currentQuestion.question}
|
| 1031 |
-
|
| 1032 |
<div id="answer-options">
|
| 1033 |
{currentQuestion.options.map((option) => (
|
| 1034 |
<div
|
| 1035 |
key={option}
|
| 1036 |
className={`answer-option ${selectedAnswer === option ? selectedAnswerColor : ''} ${selectedAnswer ? 'disabled' : ''} ${selectedAnswerColor && option === currentQuestion.correctAnswer ? 'correct' : ''}`}
|
| 1037 |
-
onClick={() => handleAnswerSelection(option)}
|
| 1038 |
>
|
| 1039 |
{option}
|
| 1040 |
</div>
|
|
@@ -1054,15 +1182,41 @@ options: [
|
|
| 1054 |
<>
|
| 1055 |
<div id="final-score-area">
|
| 1056 |
Your Final Score: {score} / {quizData.length} <br/>
|
| 1057 |
-
{score >=
|
| 1058 |
<span className="pass-message">Passed</span>
|
| 1059 |
) : (
|
| 1060 |
<span className="fail-message">Failed</span>
|
| 1061 |
)}
|
| 1062 |
</div>
|
| 1063 |
-
<
|
| 1064 |
-
|
| 1065 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1066 |
</>
|
| 1067 |
)}
|
| 1068 |
</div>
|
|
@@ -1072,4 +1226,4 @@ options: [
|
|
| 1072 |
ReactDOM.render(<QuizApp />, document.getElementById("root"));
|
| 1073 |
</script>
|
| 1074 |
</body>
|
| 1075 |
-
</html>
|
|
|
|
| 20 |
justify-content: center;
|
| 21 |
align-items: center;
|
| 22 |
min-height: 100vh;
|
| 23 |
+
background-image: url('quiz-background.jpg');
|
| 24 |
background-size: cover;
|
| 25 |
background-position: center;
|
| 26 |
}
|
| 27 |
|
| 28 |
#root {
|
| 29 |
+
background-color: rgba(255, 255, 255, 0.95);
|
| 30 |
padding: 30px;
|
| 31 |
border-radius: 12px;
|
| 32 |
+
box-shadow: 0 8px 20px rgba(0, 0, 0, 0.1);
|
| 33 |
+
width: 80%;
|
| 34 |
+
max-width: 800px;
|
| 35 |
text-align: center;
|
| 36 |
+
box-sizing: border-box;
|
| 37 |
}
|
| 38 |
|
| 39 |
h1 {
|
| 40 |
font-size: 2rem;
|
| 41 |
margin-bottom: 20px;
|
| 42 |
+
color: #3498db;
|
| 43 |
+
font-weight: 600;
|
| 44 |
}
|
| 45 |
|
| 46 |
p {
|
| 47 |
font-size: 1.1rem;
|
| 48 |
+
margin-bottom: 25px;
|
| 49 |
+
color: #555;
|
| 50 |
+
line-height: 1.7;
|
| 51 |
}
|
| 52 |
|
| 53 |
#question-area {
|
| 54 |
font-size: 1.2rem;
|
| 55 |
margin-bottom: 20px;
|
| 56 |
padding: 15px;
|
| 57 |
+
background-color: #e8f0fa;
|
| 58 |
border-radius: 8px;
|
| 59 |
+
border: 1px solid #b8c6da;
|
| 60 |
+
color: #2c3e50;
|
| 61 |
+
text-align: left;
|
| 62 |
}
|
| 63 |
|
| 64 |
#answer-options {
|
| 65 |
display: flex;
|
| 66 |
flex-direction: column;
|
| 67 |
+
align-items: stretch;
|
| 68 |
+
margin-bottom: 25px;
|
| 69 |
}
|
| 70 |
|
| 71 |
.answer-option {
|
| 72 |
padding: 15px;
|
| 73 |
+
margin-bottom: 12px;
|
| 74 |
+
background-color: #fff;
|
| 75 |
border-radius: 8px;
|
| 76 |
+
border: 1px solid #ddd;
|
| 77 |
cursor: pointer;
|
| 78 |
+
transition: background-color 0.3s ease, transform 0.2s ease;
|
| 79 |
+
font-size: 1.1rem;
|
| 80 |
+
text-align: left;
|
| 81 |
+
box-shadow: 0 2px 4px rgba(0, 0, 0, 0.05);
|
| 82 |
}
|
| 83 |
|
| 84 |
.answer-option:hover {
|
| 85 |
+
background-color: #f0f8ff;
|
| 86 |
+
transform: translateY(-2px);
|
| 87 |
+
border-color: #a6d4fa;
|
| 88 |
}
|
| 89 |
|
| 90 |
.answer-option.selected {
|
| 91 |
+
background-color: #a6d4fa;
|
| 92 |
+
border-color: #3498db;
|
| 93 |
+
color: #fff;
|
| 94 |
+
box-shadow: 0 4px 6px rgba(0, 0, 0, 0.1);
|
| 95 |
}
|
| 96 |
|
| 97 |
.answer-option.correct {
|
| 98 |
+
background-color: #86ef7d;
|
| 99 |
+
border-color: #22c55e;
|
| 100 |
color: #fff;
|
| 101 |
+
font-weight: 600;
|
| 102 |
}
|
| 103 |
|
| 104 |
.answer-option.incorrect {
|
| 105 |
+
background-color: #fca5a5;
|
| 106 |
+
border-color: #dc2626;
|
| 107 |
color: #fff;
|
| 108 |
+
font-weight: 600;
|
| 109 |
}
|
| 110 |
|
| 111 |
#result-message {
|
| 112 |
font-size: 1.2rem;
|
| 113 |
margin-bottom: 25px;
|
| 114 |
+
font-weight: 500;
|
| 115 |
}
|
| 116 |
|
| 117 |
.correct-message {
|
| 118 |
+
color: #22c55e;
|
| 119 |
}
|
| 120 |
|
| 121 |
.incorrect-message {
|
| 122 |
+
color: #dc2626;
|
| 123 |
}
|
| 124 |
|
| 125 |
+
#next-button, #skip-button, #reset-button, #review-button {
|
|
|
|
| 126 |
padding: 12px 25px;
|
| 127 |
font-size: 1.1rem;
|
| 128 |
+
margin: 5px 10px;
|
| 129 |
border-radius: 8px;
|
| 130 |
border: none;
|
| 131 |
cursor: pointer;
|
| 132 |
+
transition: background-color 0.3s ease, transform 0.2s ease, box-shadow 0.3s ease;
|
| 133 |
+
font-weight: 500;
|
| 134 |
+
box-shadow: 0 2px 5px rgba(0, 0, 0, 0.1);
|
| 135 |
}
|
| 136 |
|
| 137 |
#next-button {
|
| 138 |
+
background-color: #3498db;
|
| 139 |
color: #fff;
|
| 140 |
}
|
| 141 |
#next-button:hover {
|
| 142 |
background-color: #2980b9;
|
| 143 |
transform: translateY(-2px);
|
| 144 |
+
box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15);
|
| 145 |
}
|
| 146 |
#next-button:disabled {
|
| 147 |
+
background-color: #b8c6da;
|
| 148 |
cursor: not-allowed;
|
| 149 |
+
transform: none;
|
| 150 |
+
box-shadow: none;
|
| 151 |
}
|
| 152 |
|
| 153 |
#skip-button {
|
| 154 |
+
background-color: #f39c12;
|
| 155 |
color: #fff;
|
| 156 |
}
|
| 157 |
#skip-button:hover {
|
| 158 |
background-color: #e67e22;
|
| 159 |
transform: translateY(-2px);
|
| 160 |
+
box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15);
|
| 161 |
}
|
| 162 |
|
| 163 |
#reset-button {
|
| 164 |
+
background-color: #2ecc71;
|
| 165 |
color: #fff;
|
| 166 |
}
|
| 167 |
#reset-button:hover {
|
| 168 |
background-color: #27ae60;
|
| 169 |
transform: translateY(-2px);
|
| 170 |
+
box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15);
|
| 171 |
+
}
|
| 172 |
+
|
| 173 |
+
#review-button {
|
| 174 |
+
background-color: #8e44ad;
|
| 175 |
+
color: #fff;
|
| 176 |
+
}
|
| 177 |
+
#review-button:hover {
|
| 178 |
+
background-color: #732d91;
|
| 179 |
+
transform: translateY(-2px);
|
| 180 |
+
box-shadow: 0 4px 7px rgba(0, 0, 0, 0.15);
|
| 181 |
}
|
| 182 |
|
| 183 |
#final-score-area {
|
| 184 |
font-size: 1.5rem;
|
| 185 |
font-weight: 600;
|
| 186 |
margin-bottom: 30px;
|
| 187 |
+
color: #2c3e50;
|
| 188 |
}
|
| 189 |
|
| 190 |
.pass-message {
|
| 191 |
+
color: #22c55e;
|
| 192 |
}
|
| 193 |
|
| 194 |
.fail-message {
|
| 195 |
+
color: #dc2626;
|
| 196 |
}
|
| 197 |
|
| 198 |
+
#review-section {
|
| 199 |
+
text-align: left;
|
| 200 |
+
margin-top: 20px;
|
| 201 |
+
}
|
| 202 |
+
|
| 203 |
+
.review-item {
|
| 204 |
+
margin-bottom: 20px;
|
| 205 |
+
padding: 15px;
|
| 206 |
+
background-color: #f9f9f9;
|
| 207 |
+
border-radius: 8px;
|
| 208 |
+
border: 1px solid #ddd;
|
| 209 |
+
}
|
| 210 |
+
|
| 211 |
+
.review-question {
|
| 212 |
+
font-size: 1.2rem;
|
| 213 |
+
color: #2c3e50;
|
| 214 |
+
margin-bottom: 10px;
|
| 215 |
+
}
|
| 216 |
+
|
| 217 |
+
.review-answer {
|
| 218 |
+
font-size: 1.1rem;
|
| 219 |
+
margin-bottom: 5px;
|
| 220 |
+
}
|
| 221 |
+
|
| 222 |
+
.review-explanation {
|
| 223 |
+
font-size: 1rem;
|
| 224 |
+
color: #555;
|
| 225 |
+
margin-top: 10px;
|
| 226 |
+
}
|
| 227 |
+
|
| 228 |
+
@media (max-width: 768px) {
|
| 229 |
#root {
|
| 230 |
width: 95%;
|
| 231 |
padding: 20px;
|
|
|
|
| 239 |
font-size: 1.1rem;
|
| 240 |
padding: 12px;
|
| 241 |
}
|
| 242 |
+
#next-button, #skip-button, #reset-button, #review-button {
|
| 243 |
padding: 10px 20px;
|
| 244 |
font-size: 1rem;
|
| 245 |
}
|
|
|
|
| 251 |
}
|
| 252 |
}
|
| 253 |
|
| 254 |
+
@media (max-width: 480px) {
|
| 255 |
#root {
|
| 256 |
width: 100%;
|
| 257 |
padding: 15px;
|
| 258 |
+
border-radius: 0;
|
| 259 |
+
box-shadow: none;
|
| 260 |
}
|
| 261 |
.answer-option {
|
| 262 |
padding: 10px;
|
|
|
|
| 267 |
font-size: 1rem;
|
| 268 |
padding: 10px;
|
| 269 |
}
|
| 270 |
+
#next-button, #skip-button, #reset-button, #review-button {
|
| 271 |
padding: 10px 18px;
|
| 272 |
font-size: 0.9rem;
|
| 273 |
+
margin: 5px 5px;
|
| 274 |
}
|
| 275 |
h1 {
|
| 276 |
font-size: 1.5rem;
|
|
|
|
| 301 |
"To predict stock market trends.",
|
| 302 |
],
|
| 303 |
correctAnswer: "To understand and generate human language.",
|
| 304 |
+
explanation: "Language modeling aims to understand and generate human language by predicting the likelihood of word sequences, forming the basis for many NLP applications."
|
| 305 |
},
|
| 306 |
{
|
| 307 |
question: "Which of the following is NOT a stage in the development of language models?",
|
|
|
|
| 312 |
"Quantum language models (QLM)",
|
| 313 |
],
|
| 314 |
correctAnswer: "Quantum language models (QLM)",
|
| 315 |
+
explanation: "Quantum language models (QLM) are not a recognized stage in language model development; the progression includes SLMs, NLMs, and PLMs."
|
| 316 |
},
|
| 317 |
{
|
| 318 |
question: "What is the basic idea behind statistical language models (SLMs)?",
|
|
|
|
| 323 |
"To pre-train Transformer models.",
|
| 324 |
],
|
| 325 |
correctAnswer: "To build word prediction models based on the Markov assumption.",
|
| 326 |
+
explanation: "SLMs rely on the Markov assumption, which posits that the probability of a word depends only on a fixed number of previous words (n-grams)."
|
| 327 |
},
|
| 328 |
{
|
| 329 |
question: "What is a limitation of SLMs?",
|
|
|
|
| 334 |
"They are not used in NLP.",
|
| 335 |
],
|
| 336 |
correctAnswer: "They suffer from the curse of dimensionality.",
|
| 337 |
+
explanation: "SLMs struggle with high-dimensional data due to the curse of dimensionality, where the number of possible n-grams grows exponentially, making it hard to estimate probabilities accurately."
|
| 338 |
},
|
| 339 |
{
|
| 340 |
question: "What is the main characteristic of neural language models (NLMs)?",
|
|
|
|
| 345 |
"They use n-gram models.",
|
| 346 |
],
|
| 347 |
correctAnswer: "They characterize the probability of word sequences by neural networks.",
|
| 348 |
+
explanation: "NLMs use neural networks to model the probability of word sequences, overcoming some limitations of SLMs by learning continuous representations."
|
| 349 |
},
|
| 350 |
{
|
| 351 |
question: "What concept did the work in [1] introduce?",
|
|
|
|
| 356 |
"Pre-training and fine-tuning.",
|
| 357 |
],
|
| 358 |
correctAnswer: "Distributed representation of words.",
|
| 359 |
+
explanation: "The work in [1] refers to early neural network models (e.g., word2vec) that introduced distributed representations, allowing words to be represented as dense vectors in a continuous space."
|
| 360 |
},
|
| 361 |
{
|
| 362 |
question: "What is word2vec?",
|
|
|
|
| 367 |
"A pre-trained language model.",
|
| 368 |
],
|
| 369 |
correctAnswer: "A simplified shallow neural network for learning distributed word representations.",
|
| 370 |
+
explanation: "Word2vec is a shallow neural network designed to learn distributed word representations (word embeddings) efficiently from large text corpora."
|
| 371 |
},
|
| 372 |
{
|
| 373 |
question: "What was an early attempt at capturing context-aware word representations?",
|
| 374 |
options: ["BERT", "ELMo", "GPT-2", "word2vec"],
|
| 375 |
correctAnswer: "ELMo",
|
| 376 |
+
explanation: "ELMo (Embeddings from Language Models) was an early model that captured context-aware word representations by using bidirectional LSTMs, unlike the static embeddings of word2vec."
|
| 377 |
},
|
| 378 |
{
|
| 379 |
question: "Which architecture is highly parallelizable and used in BERT?",
|
| 380 |
options: ["RNN", "LSTM", "Transformer", "MLP"],
|
| 381 |
correctAnswer: "Transformer",
|
| 382 |
+
explanation: "BERT uses the Transformer architecture, which is highly parallelizable due to its self-attention mechanism, unlike sequential models like RNNs or LSTMs."
|
| 383 |
},
|
| 384 |
{
|
| 385 |
question: "What is a key feature of BERT?",
|
|
|
|
| 390 |
"It predicts the next word based on the most recent context.",
|
| 391 |
],
|
| 392 |
correctAnswer: "It pre-trains bidirectional language models.",
|
| 393 |
+
explanation: "BERT’s key feature is its bidirectional pre-training, allowing it to consider both left and right context for each word, unlike unidirectional models."
|
| 394 |
},
|
| 395 |
{
|
| 396 |
question: "What paradigm did BERT inspire?",
|
|
|
|
| 401 |
"Distributed representation learning.",
|
| 402 |
],
|
| 403 |
correctAnswer: "Pre-training and fine-tuning.",
|
| 404 |
+
explanation: "BERT popularized the pre-training and fine-tuning paradigm, where a model is pre-trained on a large corpus and then fine-tuned for specific tasks."
|
| 405 |
},
|
| 406 |
{
|
| 407 |
question: "What do researchers find about scaling PLMs?",
|
|
|
|
| 412 |
"It only works for small models.",
|
| 413 |
],
|
| 414 |
correctAnswer: "It often leads to an improved model capacity.",
|
| 415 |
+
explanation: "Scaling pre-trained language models (PLMs) by increasing parameters and data often improves their capacity to handle complex tasks, as seen in models like GPT-3."
|
| 416 |
},
|
| 417 |
{
|
| 418 |
question: "What is a characteristic of large-sized PLMs (LLMs)?",
|
|
|
|
| 423 |
"They are less complex than SLMs.",
|
| 424 |
],
|
| 425 |
correctAnswer: "They display different behaviors from smaller PLMs.",
|
| 426 |
+
explanation: "Large-sized PLMs (LLMs) exhibit emergent behaviors, such as few-shot learning, that smaller PLMs like BERT typically do not show."
|
| 427 |
},
|
| 428 |
{
|
| 429 |
question: "What is an example of an emergent ability in LLMs?",
|
|
|
|
| 434 |
"Using only n-gram models.",
|
| 435 |
],
|
| 436 |
correctAnswer: "Solving few-shot tasks through in-context learning.",
|
| 437 |
+
explanation: "An emergent ability in LLMs is solving few-shot tasks via in-context learning, where the model adapts to new tasks with just a few examples provided in the prompt."
|
| 438 |
},
|
| 439 |
+
{
|
| 440 |
question: "What does the term 'LLM' stand for?",
|
| 441 |
options: [
|
| 442 |
"Limited Language Model",
|
|
|
|
| 445 |
"Logical Language Model",
|
| 446 |
],
|
| 447 |
correctAnswer: "Large Language Model",
|
| 448 |
+
explanation: "LLM stands for Large Language Model, referring to models with billions of parameters trained on massive datasets."
|
| 449 |
},
|
| 450 |
{
|
| 451 |
question: "Which of the following is a remarkable application of LLMs?",
|
| 452 |
options: ["ELMo", "ChatGPT", "word2vec", "SLM"],
|
| 453 |
correctAnswer: "ChatGPT",
|
| 454 |
+
explanation: "ChatGPT, built on the GPT architecture, is a remarkable LLM application known for its conversational abilities and widespread use."
|
| 455 |
},
|
| 456 |
{
|
| 457 |
question: "What is a key focus of the latest language models (e.g., GPT-4)?",
|
|
|
|
| 462 |
"Using only statistical methods.",
|
| 463 |
],
|
| 464 |
correctAnswer: "Complex task solving.",
|
| 465 |
+
explanation: "Latest models like GPT-4 focus on solving complex tasks, leveraging their scale and training to handle diverse, intricate problems."
|
| 466 |
},
|
| 467 |
{
|
| 468 |
question: "How do LLMs differ from small PLMs in accessing them?",
|
|
|
|
| 473 |
"They are accessed using n-gram models.",
|
| 474 |
],
|
| 475 |
correctAnswer: "They are accessed through the prompting interface.",
|
| 476 |
+
explanation: "LLMs are typically accessed via prompting, where users provide instructions or examples in natural language, unlike smaller PLMs that often require fine-tuning."
|
| 477 |
},
|
| 478 |
{
|
| 479 |
question: "What is a challenge in developing LLMs?",
|
|
|
|
| 484 |
"It is very costly to train them due to huge demand for computation resources.",
|
| 485 |
],
|
| 486 |
correctAnswer: "It is very costly to train them due to huge demand for computation resources.",
|
| 487 |
+
explanation: "Training LLMs requires vast computational resources, making it expensive and often limiting development to well-funded organizations."
|
| 488 |
},
|
| 489 |
{
|
| 490 |
question: "What is a potential issue with LLMs despite their capacities?",
|
|
|
|
| 495 |
"They do not require effective control approaches.",
|
| 496 |
],
|
| 497 |
correctAnswer: "They are likely to produce toxic, fictitious, or harmful content.",
|
| 498 |
+
explanation: "Despite their capabilities, LLMs can generate toxic or false content due to biases in training data or lack of perfect alignment with human values."
|
| 499 |
},
|
| 500 |
{
|
| 501 |
question: "What are the four major aspects of LLMs covered in the survey?",
|
|
|
|
| 506 |
"Input, processing, output, and storage.",
|
| 507 |
],
|
| 508 |
correctAnswer: "Pre-training, adaptation, utilization, and capacity evaluation.",
|
| 509 |
+
explanation: "The survey covers pre-training (initial training), adaptation (tuning), utilization (application), and capacity evaluation (performance assessment) as key aspects of LLMs."
|
| 510 |
},
|
| 511 |
{
|
| 512 |
question: "What is the typical parameter size of LLMs?",
|
|
|
|
| 517 |
"Less than 100 parameters.",
|
| 518 |
],
|
| 519 |
correctAnswer: "Billions or hundreds of billions of parameters.",
|
| 520 |
+
explanation: "LLMs typically have billions or hundreds of billions of parameters, enabling their vast capacity, unlike smaller models with millions."
|
| 521 |
},
|
| 522 |
{
|
| 523 |
question: "On what type of data are LLMs typically trained?",
|
|
|
|
| 528 |
"Audio data only.",
|
| 529 |
],
|
| 530 |
correctAnswer: "Massive text data.",
|
| 531 |
+
explanation: "LLMs are trained on massive text corpora, such as web texts, books, and articles, to capture broad language patterns."
|
| 532 |
},
|
| 533 |
{
|
| 534 |
question: "What strong capacity do LLMs exhibit?",
|
|
|
|
| 539 |
"Performing only simple calculations.",
|
| 540 |
],
|
| 541 |
correctAnswer: "Solving complex tasks via text generation.",
|
| 542 |
+
explanation: "LLMs excel at solving complex tasks by generating text, leveraging their understanding of language and context."
|
| 543 |
},
|
| 544 |
{
|
| 545 |
question: "What does the survey introduce as basic background for LLMs?",
|
|
|
|
| 550 |
"Only scaling laws.",
|
| 551 |
],
|
| 552 |
correctAnswer: "Scaling laws, emergent abilities, and key techniques.",
|
| 553 |
+
explanation: "The survey provides background on scaling laws (performance vs. size), emergent abilities (e.g., few-shot learning), and key techniques (e.g., Transformers)."
|
| 554 |
},
|
| 555 |
{
|
| 556 |
question: "What is the key to understanding the development of language models in research history?",
|
|
|
|
| 561 |
"Ignoring the evolution of model capacities.",
|
| 562 |
],
|
| 563 |
correctAnswer: "The leap from language modeling to task solving.",
|
| 564 |
+
explanation: "The shift from merely modeling language (predicting words) to solving tasks (e.g., reasoning, Q&A) marks a pivotal development in language models."
|
| 565 |
},
|
| 566 |
{
|
| 567 |
question: "Which of the following is NOT a characteristic of LLMs?",
|
|
|
|
| 572 |
"Using Transformer language models.",
|
| 573 |
],
|
| 574 |
correctAnswer: "Training on small datasets.",
|
| 575 |
+
explanation: "LLMs are characterized by training on massive datasets, not small ones, which enables their strong performance."
|
| 576 |
},
|
| 577 |
{
|
| 578 |
question: "What is a major difference between LLMs and previous smaller PLMs?",
|
|
|
|
| 583 |
"LLMs do not require large-scale data.",
|
| 584 |
],
|
| 585 |
correctAnswer: "LLMs display surprising emergent abilities.",
|
| 586 |
+
explanation: "LLMs show emergent abilities like in-context learning, which smaller PLMs typically lack due to their scale and training."
|
| 587 |
},
|
| 588 |
{
|
| 589 |
question: "How has the research paradigm shifted towards the use of LLMs in NLP?",
|
|
|
|
| 594 |
"Ignoring pre-training and fine-tuning.",
|
| 595 |
],
|
| 596 |
correctAnswer: "LLMs serve as a general-purpose language task solver.",
|
| 597 |
+
explanation: "The paradigm has shifted to using LLMs as general-purpose solvers for various NLP tasks via prompting, rather than task-specific fine-tuning."
|
| 598 |
},
|
| 599 |
{
|
| 600 |
question: "What is a challenge that traditional search engines face with the advent of LLMs?",
|
|
|
|
| 605 |
"Easier data processing.",
|
| 606 |
],
|
| 607 |
correctAnswer: "New information seeking way through AI chatbots.",
|
| 608 |
+
explanation: "LLM-powered chatbots offer a conversational way to seek information, challenging the keyword-based approach of traditional search engines."
|
| 609 |
},
|
| 610 |
{
|
| 611 |
question: "What is a current trend in computer vision (CV) research related to LLMs?",
|
|
|
|
| 616 |
"Focusing solely on image classification.",
|
| 617 |
],
|
| 618 |
correctAnswer: "Developing ChatGPT-like vision-language models.",
|
| 619 |
+
explanation: "CV research is trending towards multimodal models that combine vision and language, inspired by ChatGPT’s success."
|
| 620 |
},
|
| 621 |
{
|
| 622 |
question: "What is a potential impact of the new wave of LLM technology?",
|
|
|
|
| 627 |
"Slower development of AI algorithms.",
|
| 628 |
],
|
| 629 |
correctAnswer: "A prosperous ecosystem of real-world applications based on LLMs.",
|
| 630 |
+
explanation: "LLMs are fostering a wide range of real-world applications, from chatbots to automation, due to their versatility."
|
| 631 |
},
|
| 632 |
{
|
| 633 |
question: "What is one of the mysterious aspects of LLMs?",
|
|
|
|
| 638 |
"Why they only work on small datasets.",
|
| 639 |
],
|
| 640 |
correctAnswer: "Why emergent abilities occur in LLMs.",
|
| 641 |
+
explanation: "The emergence of abilities like few-shot learning in LLMs is not fully understood, making it a mysterious aspect of their behavior."
|
| 642 |
},
|
| 643 |
{
|
| 644 |
question: "Why is it difficult for the research community to train capable LLMs?",
|
|
|
|
| 649 |
"Because LLMs are mainly trained by academia.",
|
| 650 |
],
|
| 651 |
correctAnswer: "Because it is very costly to carry out repetitive studies.",
|
| 652 |
+
explanation: "Training LLMs requires extensive computational resources, making repetitive studies costly and limiting academic research."
|
| 653 |
},
|
| 654 |
{
|
| 655 |
question: "What is a challenge in aligning LLMs?",
|
|
|
|
| 660 |
"Reducing their computational costs.",
|
| 661 |
],
|
| 662 |
correctAnswer: "Aligning them with human values or preferences.",
|
| 663 |
+
explanation: "Aligning LLMs with human values is challenging due to biases in data and the complexity of defining universal preferences."
|
| 664 |
},
|
| 665 |
{
|
| 666 |
question: "What does the paper 'Planning for AGI and beyond' discuss?",
|
|
|
|
| 671 |
"Only short-term plans for AI development.",
|
| 672 |
],
|
| 673 |
correctAnswer: "Short-term and long-term plans to approach AGI.",
|
| 674 |
+
explanation: "The paper outlines strategies for developing Artificial General Intelligence (AGI), covering both immediate and future steps."
|
| 675 |
},
|
| 676 |
{
|
| 677 |
question: "What is a recent argument about GPT-4?",
|
|
|
|
| 682 |
"It can only perform simple tasks.",
|
| 683 |
],
|
| 684 |
correctAnswer: "It might be considered an early version of an AGI system.",
|
| 685 |
+
explanation: "Some argue GPT-4’s broad capabilities suggest it could be an early AGI, though it lacks full general intelligence."
|
| 686 |
},
|
| 687 |
{
|
| 688 |
question: "How is Microsoft 365 being empowered?",
|
|
|
|
| 693 |
"By using only statistical models.",
|
| 694 |
],
|
| 695 |
correctAnswer: "By LLMs to automate office work.",
|
| 696 |
+
explanation: "Microsoft 365 integrates LLMs (e.g., via Copilot) to automate tasks like writing, summarizing, and data analysis."
|
| 697 |
},
|
| 698 |
{
|
| 699 |
+
question: "What is a suggestion for choosing layer normalization in LLMs?",
|
| 700 |
+
options: ["Post RMSNorm", "Pre RMSNorm", "Post LN", "No Normalization"],
|
| 701 |
+
correctAnswer: "Pre RMSNorm",
|
| 702 |
+
explanation: "Pre RMSNorm (Root Mean Square Normalization before layers) is suggested for LLMs due to its stability and performance benefits."
|
| 703 |
},
|
| 704 |
{
|
| 705 |
+
question: "Which activation function is recommended for stronger generalization and training stability?",
|
| 706 |
+
options: ["ReLU", "Sigmoid", "SwiGLU", "Tanh"],
|
| 707 |
+
correctAnswer: "SwiGLU",
|
| 708 |
+
explanation: "SwiGLU (Swish-Gated Linear Unit) is recommended for LLMs as it improves generalization and training stability over traditional functions like ReLU."
|
| 709 |
},
|
| 710 |
{
|
| 711 |
+
question: "Which position embedding is considered a better choice for LLMs?",
|
| 712 |
+
options: ["Absolute Positional Embedding", "Relative Positional Embedding", "RoPE", "Sinusoidal Positional Encoding"],
|
| 713 |
+
correctAnswer: "RoPE",
|
| 714 |
+
explanation: "RoPE (Rotary Position Embedding) is favored in LLMs for its ability to encode relative positions efficiently and scale with sequence length."
|
| 715 |
},
|
| 716 |
{
|
| 717 |
question: "What is the primary role of pre-training in LLMs?",
|
| 718 |
+
options: [
|
| 719 |
"To fine-tune models for specific tasks.",
|
| 720 |
"To encode general knowledge from large-scale corpus.",
|
| 721 |
"To reduce the size of the model.",
|
| 722 |
"To improve inference speed.",
|
| 723 |
],
|
| 724 |
correctAnswer: "To encode general knowledge from large-scale corpus.",
|
| 725 |
+
explanation: "Pre-training encodes general knowledge from vast text corpora into LLMs, providing a foundation for later task-specific adaptation."
|
| 726 |
+
},
|
| 727 |
{
|
| 728 |
question: "What type of data is included in the arXiv Dataset?",
|
| 729 |
options: [
|
|
|
|
| 733 |
"Image and video data.",
|
| 734 |
],
|
| 735 |
correctAnswer: "Scientific publication data.",
|
| 736 |
+
explanation: "The arXiv Dataset contains scientific publication data, primarily research papers, used for training models on academic content."
|
| 737 |
},
|
| 738 |
{
|
| 739 |
question: "What is the approximate size of the peS2o dataset?",
|
| 740 |
options: ["42MB", "42GB", "42TB", "42B tokens"],
|
| 741 |
correctAnswer: "42B tokens",
|
| 742 |
+
explanation: "The peS2o dataset is approximately 42 billion tokens, a massive text corpus used for training LLMs."
|
| 743 |
},
|
| 744 |
{
|
| 745 |
question: "What is a characteristic of the articles in Wikipedia?",
|
|
|
|
| 750 |
"They cover only a narrow range of topics.",
|
| 751 |
],
|
| 752 |
correctAnswer: "They are composed in an expository style with references.",
|
| 753 |
+
explanation: "Wikipedia articles are written in an expository style, providing detailed explanations with references, making them a valuable training resource."
|
| 754 |
},
|
| 755 |
{
|
| 756 |
question: "What is a technique used to improve memory efficiency and throughput of deployed LLMs?",
|
| 757 |
options: ["Data Parallelism", "Tensor Parallelism", "Pipeline Parallelism", "PagedAttention"],
|
| 758 |
correctAnswer: "PagedAttention",
|
| 759 |
+
explanation: "PagedAttention improves memory efficiency and throughput in LLMs by managing key-value caches more effectively during inference."
|
| 760 |
},
|
| 761 |
{
|
| 762 |
question: "How does PagedAttention partition sequences?",
|
|
|
|
| 767 |
"Into overlapping segments.",
|
| 768 |
],
|
| 769 |
correctAnswer: "Into subsequences.",
|
| 770 |
+
explanation: "PagedAttention partitions sequences into subsequences, allowing efficient memory management by processing them in blocks."
|
| 771 |
},
|
| 772 |
{
|
| 773 |
question: "What is the benefit of using PagedAttention?",
|
|
|
|
| 778 |
"Improves model accuracy.",
|
| 779 |
],
|
| 780 |
correctAnswer: "Increases GPU utilization and enables efficient memory sharing.",
|
| 781 |
+
explanation: "PagedAttention boosts GPU utilization and memory sharing, optimizing resource use during LLM inference."
|
| 782 |
},
|
| 783 |
{
|
| 784 |
question: "What type of floating-point number was predominantly used for pre-training in previous PLMs like BERT?",
|
| 785 |
options: ["FP16", "BF16", "FP32", "INT8"],
|
| 786 |
correctAnswer: "FP32",
|
| 787 |
+
explanation: "FP32 (32-bit floating-point) was commonly used in earlier PLMs like BERT for its high precision during pre-training."
|
| 788 |
},
|
| 789 |
{
|
| 790 |
question: "Why have some studies started to use FP16 for pre-training LLMs?",
|
|
|
|
| 795 |
"To avoid the loss of computational accuracy.",
|
| 796 |
],
|
| 797 |
correctAnswer: "To reduce memory usage and communication overhead.",
|
| 798 |
+
explanation: "FP16 (16-bit floating-point) reduces memory usage and communication overhead, making pre-training LLMs more efficient despite lower precision."
|
| 799 |
},
|
| 800 |
{
|
| 801 |
question: "What is a potential issue with using FP16 for training?",
|
|
|
|
| 806 |
"Improved model performance.",
|
| 807 |
],
|
| 808 |
correctAnswer: "Loss of computational accuracy.",
|
| 809 |
+
explanation: "FP16’s lower precision can lead to a loss of computational accuracy, potentially affecting model quality during training."
|
| 810 |
},
|
| 811 |
{
|
| 812 |
question: "What is BF16?",
|
|
|
|
| 817 |
"A method for parallel training.",
|
| 818 |
],
|
| 819 |
correctAnswer: "Brain Floating Point, an alternative to FP16.",
|
| 820 |
+
explanation: "BF16 (Brain Floating Point) is a 16-bit format developed by Google, offering a balance between FP16’s efficiency and FP32’s precision."
|
| 821 |
},
|
| 822 |
{
|
| 823 |
question: "How does BF16 compare to FP16 in terms of representation accuracy for pre-training?",
|
|
|
|
| 828 |
"BF16 is not suitable for pre-training.",
|
| 829 |
],
|
| 830 |
correctAnswer: "BF16 generally performs better than FP16.",
|
| 831 |
+
explanation: "BF16 provides better representation accuracy than FP16 due to its wider dynamic range, making it more suitable for pre-training LLMs."
|
| 832 |
},
|
| 833 |
{
|
| 834 |
question: "Which training technique is often used jointly with 3D parallelism to improve training throughput?",
|
|
|
|
| 839 |
"PagedAttention.",
|
| 840 |
],
|
| 841 |
correctAnswer: "Mixed precision training.",
|
| 842 |
+
explanation: "Mixed precision training, combining FP16/BF16 with FP32, is used with 3D parallelism (data, tensor, pipeline) to boost LLM training throughput."
|
| 843 |
},
|
| 844 |
{
|
| 845 |
question: "What type of parallelism was used to train BLOOM on 384 A100 GPUs?",
|
|
|
|
| 850 |
"Only pipeline parallelism.",
|
| 851 |
],
|
| 852 |
correctAnswer: "8-way data parallelism, 4-way tensor parallelism, and 12-way pipeline parallelism.",
|
| 853 |
+
explanation: "BLOOM used a combination of 8-way data, 4-way tensor, and 12-way pipeline parallelism to efficiently train on 384 A100 GPUs."
|
| 854 |
},
|
| 855 |
{
|
| 856 |
question: "What is the primary approach to using LLMs after pre-training or adaptation tuning?",
|
|
|
|
| 861 |
"Using only n-gram models.",
|
| 862 |
],
|
| 863 |
correctAnswer: "Designing suitable prompting strategies.",
|
| 864 |
+
explanation: "Post-pre-training, LLMs are primarily used via prompting strategies, where carefully crafted inputs elicit desired outputs."
|
| 865 |
},
|
| 866 |
{
|
| 867 |
question: "What is a representative prompting method discussed in the text?",
|
| 868 |
options: ["Fine-tuning", "Backpropagation", "In-context learning", "Random search"],
|
| 869 |
correctAnswer: "In-context learning",
|
| 870 |
+
explanation: "In-context learning is a key prompting method where LLMs learn tasks from examples provided in the input prompt."
|
| 871 |
},
|
| 872 |
{
|
| 873 |
question: "What does in-context learning involve?",
|
|
|
|
| 878 |
"Employing only manual creation of prompts.",
|
| 879 |
],
|
| 880 |
correctAnswer: "Formulating task description and demonstrations in natural language text.",
|
| 881 |
+
explanation: "In-context learning involves providing a task description and examples in natural language within the prompt to guide the LLM."
|
| 882 |
},
|
| 883 |
{
|
| 884 |
question: "What is the process of manually creating a suitable prompt also called?",
|
| 885 |
options: ["Automatic prompt optimization", "Prompt engineering", "Prompt tuning", "Prompt generation"],
|
| 886 |
correctAnswer: "Prompt engineering",
|
| 887 |
+
explanation: "Prompt engineering refers to the manual design of prompts to effectively leverage LLMs’ capabilities for specific tasks."
|
| 888 |
},
|
| 889 |
{
|
| 890 |
question: "What is the impact of a well-designed prompt on LLMs?",
|
|
|
|
| 895 |
"It makes LLMs generate random outputs.",
|
| 896 |
],
|
| 897 |
correctAnswer: "It is very helpful to elicit the abilities of LLMs.",
|
| 898 |
+
explanation: "A well-designed prompt significantly enhances an LLM’s ability to perform tasks by providing clear context and instructions."
|
| 899 |
},
|
| 900 |
{
|
| 901 |
question: "What type of data is considered well-organized with algorithmic logic and programming flow?",
|
| 902 |
options: ["Natural language text", "Image data", "Code data", "Audio data"],
|
| 903 |
correctAnswer: "Code data",
|
| 904 |
+
explanation: "Code data is structured with algorithmic logic and programming flow, making it distinct from unstructured natural language text."
|
| 905 |
},
|
| 906 |
{
|
| 907 |
question: "What ability do models trained on code show?",
|
| 908 |
options: ["Weak reasoning ability", "Strong reasoning ability", "No reasoning ability", "Only language generation ability"],
|
| 909 |
correctAnswer: "Strong reasoning ability",
|
| 910 |
+
explanation: "Models trained on code exhibit strong reasoning ability due to the logical and structured nature of programming data."
|
| 911 |
},
|
| 912 |
{
|
| 913 |
question: "What is a hypothesis regarding code data and LLMs' reasoning performance?",
|
|
|
|
| 918 |
"Code data is only useful for code generation.",
|
| 919 |
],
|
| 920 |
correctAnswer: "Code data may be useful to improve the reasoning performance of LLMs.",
|
| 921 |
+
explanation: "It’s hypothesized that code data’s logical structure could enhance LLMs’ reasoning skills beyond just code generation."
|
| 922 |
},
|
| 923 |
{
|
| 924 |
question: "What is a characteristic of LLMs' text generation quality?",
|
|
|
|
| 929 |
"It cannot be evaluated.",
|
| 930 |
],
|
| 931 |
correctAnswer: "It is comparable to human-written texts.",
|
| 932 |
+
explanation: "LLMs generate text that is often comparable to human-written content in coherence and quality, a testament to their training scale."
|
| 933 |
},
|
| 934 |
{
|
| 935 |
question: "How can LLMs be used in the context of generation evaluation?",
|
|
|
|
| 940 |
"To perform only statistical analysis.",
|
| 941 |
],
|
| 942 |
correctAnswer: "As language generation evaluators.",
|
| 943 |
+
explanation: "LLMs can evaluate generated text by assessing its quality, coherence, or relevance, supplementing human judgment."
|
| 944 |
},
|
| 945 |
{
|
| 946 |
question: "What is a limitation of LLMs in specialized generation?",
|
|
|
|
| 951 |
"They are only good at generating code.",
|
| 952 |
],
|
| 953 |
correctAnswer: "They have learned general language patterns but underperform in specialized generation.",
|
| 954 |
+
explanation: "LLMs excel in general language but may struggle with highly specialized domains due to limited domain-specific training data."
|
| 955 |
},
|
| 956 |
{
|
| 957 |
question: "What is a common approach to enhancing LLMs' factual knowledge?",
|
|
|
|
| 962 |
"Ignoring up-to-date information.",
|
| 963 |
],
|
| 964 |
correctAnswer: "Incorporating extracted relevant information into the context.",
|
| 965 |
+
explanation: "Enhancing LLMs’ factual knowledge often involves adding relevant external information (e.g., via retrieval-augmented generation) to the context."
|
| 966 |
},
|
| 967 |
{
|
| 968 |
question: "What is a finding about smaller models with instruction tuning compared to larger models without it?",
|
|
|
|
| 973 |
"Larger models are always better.",
|
| 974 |
],
|
| 975 |
correctAnswer: "Smaller models can perform better.",
|
| 976 |
+
explanation: "Smaller models with instruction tuning can outperform larger untuned models by being more aligned with specific tasks."
|
| 977 |
},
|
| 978 |
{
|
| 979 |
question: "What does instruction tuning enable LLMs to do?",
|
|
|
|
| 984 |
"Perform only without demonstrations.",
|
| 985 |
],
|
| 986 |
correctAnswer: "Follow human instructions to perform specific tasks, even on unseen tasks.",
|
| 987 |
+
explanation: "Instruction tuning allows LLMs to generalize to unseen tasks by following human instructions provided in natural language."
|
| 988 |
},
|
| 989 |
{
|
| 990 |
question: "What is a benefit of instruction tuning?",
|
|
|
|
| 995 |
"It does not improve model performance.",
|
| 996 |
],
|
| 997 |
correctAnswer: "It is much less costly than pre-training.",
|
| 998 |
+
explanation: "Instruction tuning is less resource-intensive than pre-training, requiring only a smaller dataset of instructions to adapt the model."
|
| 999 |
},
|
| 1000 |
{
|
| 1001 |
question: "What is the primary goal of instruction tuning?",
|
|
|
|
| 1006 |
"To generate random text",
|
| 1007 |
],
|
| 1008 |
correctAnswer: "To make LLMs better at following instructions",
|
| 1009 |
+
explanation: "The main goal of instruction tuning is to improve LLMs’ ability to accurately follow human instructions for various tasks."
|
| 1010 |
},
|
| 1011 |
{
|
| 1012 |
question: "What kind of tasks does instruction tuning help LLMs perform?",
|
|
|
|
| 1017 |
"No specific tasks",
|
| 1018 |
],
|
| 1019 |
correctAnswer: "Specific tasks without demonstrations",
|
| 1020 |
+
explanation: "Instruction tuning enables LLMs to perform specific tasks based solely on instructions, without needing example demonstrations."
|
| 1021 |
},
|
| 1022 |
{
|
| 1023 |
question: "How does instruction tuning affect LLMs' ability to follow instructions?",
|
|
|
|
| 1028 |
"It makes them ignore instructions",
|
| 1029 |
],
|
| 1030 |
correctAnswer: "It enhances their ability",
|
| 1031 |
+
explanation: "Instruction tuning enhances LLMs’ capability to interpret and act on human instructions effectively."
|
| 1032 |
},
|
| 1033 |
{
|
| 1034 |
question: "What have a large number of studies confirmed about instruction tuning?",
|
|
|
|
| 1039 |
"It decreases performance",
|
| 1040 |
],
|
| 1041 |
correctAnswer: "It achieves superior performance on seen and unseen tasks",
|
| 1042 |
+
explanation: "Studies show instruction tuning boosts LLM performance on both familiar (seen) and new (unseen) tasks."
|
| 1043 |
},
|
| 1044 |
{
|
| 1045 |
question: "What is a key aspect of high-quality long CoT data curation?",
|
|
|
|
| 1050 |
"Manual data collection only",
|
| 1051 |
],
|
| 1052 |
correctAnswer: "Using open models or APIs for data synthesis",
|
| 1053 |
+
explanation: "High-quality Chain-of-Thought (CoT) data is often curated using open models or APIs to synthesize detailed reasoning steps."
|
| 1054 |
},
|
| 1055 |
{
|
| 1056 |
question: "What is the basic idea behind creating long CoT response data?",
|
|
|
|
| 1061 |
"Ignoring the prompt structure",
|
| 1062 |
],
|
| 1063 |
correctAnswer: "Feeding prompts into teacher models",
|
| 1064 |
+
explanation: "Long CoT response data is created by feeding prompts into teacher models to generate step-by-step reasoning responses."
|
| 1065 |
},
|
| 1066 |
{
|
| 1067 |
question: "What is the finding about smaller models with instruction tuning compared to larger models without fine-tuning?",
|
|
|
|
| 1072 |
"Larger models are always better.",
|
| 1073 |
],
|
| 1074 |
correctAnswer: "Smaller models can perform better.",
|
| 1075 |
+
explanation: "Research indicates that smaller, instruction-tuned models can outperform larger models without tuning due to better task alignment."
|
| 1076 |
},
|
| 1077 |
];
|
| 1078 |
|
|
|
|
| 1084 |
const [quizEnd, setQuizEnd] = React.useState(false);
|
| 1085 |
const [selectedAnswerColor, setSelectedAnswerColor] = React.useState('');
|
| 1086 |
const [currentQuestionNumber, setCurrentQuestionNumber] = React.useState(1);
|
| 1087 |
+
const [incorrectAnswers, setIncorrectAnswers] = React.useState([]);
|
| 1088 |
+
const [showReview, setShowReview] = React.useState(false);
|
| 1089 |
|
| 1090 |
const currentQuestion = quizData[currentQuestionIndex];
|
| 1091 |
const totalQuestions = quizData.length;
|
| 1092 |
|
|
|
|
| 1093 |
const handleAnswerSelection = (answer) => {
|
| 1094 |
setSelectedAnswer(answer);
|
| 1095 |
if (answer === currentQuestion.correctAnswer) {
|
|
|
|
| 1099 |
} else {
|
| 1100 |
setMessage('Incorrect!');
|
| 1101 |
setSelectedAnswerColor('incorrect');
|
| 1102 |
+
setIncorrectAnswers([...incorrectAnswers, {
|
| 1103 |
+
question: currentQuestion.question,
|
| 1104 |
+
selectedAnswer: answer,
|
| 1105 |
+
correctAnswer: currentQuestion.correctAnswer,
|
| 1106 |
+
explanation: currentQuestion.explanation
|
| 1107 |
+
}]);
|
| 1108 |
}
|
| 1109 |
};
|
| 1110 |
|
|
|
|
| 1140 |
setQuizEnd(false);
|
| 1141 |
setSelectedAnswerColor('');
|
| 1142 |
setCurrentQuestionNumber(1);
|
| 1143 |
+
setIncorrectAnswers([]);
|
| 1144 |
+
setShowReview(false);
|
| 1145 |
+
};
|
| 1146 |
+
|
| 1147 |
+
const handleReviewIncorrect = () => {
|
| 1148 |
+
setShowReview(true);
|
| 1149 |
};
|
| 1150 |
|
| 1151 |
return (
|
|
|
|
| 1156 |
<>
|
| 1157 |
<div id="question-area">
|
| 1158 |
Question {currentQuestionNumber}/{totalQuestions}: {currentQuestion.question}
|
| 1159 |
+
</div>
|
| 1160 |
<div id="answer-options">
|
| 1161 |
{currentQuestion.options.map((option) => (
|
| 1162 |
<div
|
| 1163 |
key={option}
|
| 1164 |
className={`answer-option ${selectedAnswer === option ? selectedAnswerColor : ''} ${selectedAnswer ? 'disabled' : ''} ${selectedAnswerColor && option === currentQuestion.correctAnswer ? 'correct' : ''}`}
|
| 1165 |
+
onClick={() => !selectedAnswer && handleAnswerSelection(option)}
|
| 1166 |
>
|
| 1167 |
{option}
|
| 1168 |
</div>
|
|
|
|
| 1182 |
<>
|
| 1183 |
<div id="final-score-area">
|
| 1184 |
Your Final Score: {score} / {quizData.length} <br/>
|
| 1185 |
+
{score >= Math.ceil(quizData.length * 0.8) ? (
|
| 1186 |
<span className="pass-message">Passed</span>
|
| 1187 |
) : (
|
| 1188 |
<span className="fail-message">Failed</span>
|
| 1189 |
)}
|
| 1190 |
</div>
|
| 1191 |
+
<div style={{display: 'flex', justifyContent: 'center'}}>
|
| 1192 |
+
<button id="reset-button" onClick={handleResetQuiz}>
|
| 1193 |
+
Play Again
|
| 1194 |
+
</button>
|
| 1195 |
+
{incorrectAnswers.length > 0 && (
|
| 1196 |
+
<button id="review-button" onClick={handleReviewIncorrect}>
|
| 1197 |
+
Review Incorrect Answers
|
| 1198 |
+
</button>
|
| 1199 |
+
)}
|
| 1200 |
+
</div>
|
| 1201 |
+
{showReview && (
|
| 1202 |
+
<div id="review-section">
|
| 1203 |
+
<h2>Review of Incorrect Answers</h2>
|
| 1204 |
+
{incorrectAnswers.map((item, index) => (
|
| 1205 |
+
<div key={index} className="review-item">
|
| 1206 |
+
<div className="review-question">{item.question}</div>
|
| 1207 |
+
<div className="review-answer">
|
| 1208 |
+
<strong>Your Answer:</strong> {item.selectedAnswer} <span className="incorrect-message">(Incorrect)</span>
|
| 1209 |
+
</div>
|
| 1210 |
+
<div className="review-answer">
|
| 1211 |
+
<strong>Correct Answer:</strong> {item.correctAnswer} <span className="correct-message">(Correct)</span>
|
| 1212 |
+
</div>
|
| 1213 |
+
<div className="review-explanation">
|
| 1214 |
+
<strong>Explanation:</strong> {item.explanation}
|
| 1215 |
+
</div>
|
| 1216 |
+
</div>
|
| 1217 |
+
))}
|
| 1218 |
+
</div>
|
| 1219 |
+
)}
|
| 1220 |
</>
|
| 1221 |
)}
|
| 1222 |
</div>
|
|
|
|
| 1226 |
ReactDOM.render(<QuizApp />, document.getElementById("root"));
|
| 1227 |
</script>
|
| 1228 |
</body>
|
| 1229 |
+
</html>
|