File size: 11,261 Bytes
0c08f5a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
WEBVTT
X-TIMESTAMP-MAP=LOCAL:00:00:00.000,MPEGTS:144533

1
00:00:00.033 --> 00:00:03.303
We're going to wrap up the whole course
with these explanations.

2
00:00:03.536 --> 00:00:07.307
If you have followed the previous lab,
I've quickly mentioned,

3
00:00:07.874 --> 00:00:11.644
the notion of emergent features
for large language models.

4
00:00:12.045 --> 00:00:15.015
And this is part of
like one of the biggest challenges

5
00:00:15.015 --> 00:00:17.584
when when it comes to quantizing
large language models.

6
00:00:17.584 --> 00:00:21.688
Once in the open source community,
we had more and more large language

7
00:00:21.688 --> 00:00:26.688
models such as OPT the opened
pre-trained transformers from Facebook.

8
00:00:27.260 --> 00:00:32.260
In 2022, researchers started to directly
dive into the capabilities of the model,

9
00:00:33.166 --> 00:00:37.637
and they discovered
some so-called emergent features at scale.

10
00:00:37.771 --> 00:00:40.440
What do we mean
exactly by emergent features?

11
00:00:40.440 --> 00:00:44.911
Simply, some characteristics
or features that appear at scale

12
00:00:45.445 --> 00:00:46.846
so when the model is large.

13
00:00:46.846 --> 00:00:50.216
So it turns out that for some models
that scale

14
00:00:50.984 --> 00:00:54.054
the features predicted by the model,

15
00:00:54.788 --> 00:00:57.690
meaning the magnitude of the hidden states

16
00:00:57.690 --> 00:01:01.161
started to get large,
thus making the classic quantization

17
00:01:01.161 --> 00:01:04.531
schemes quite obsolete,
which led to, you know, classic,

18
00:01:05.198 --> 00:01:07.567
linear quantization algorithms,

19
00:01:07.567 --> 00:01:10.203
just failing on those models.

20
00:01:10.203 --> 00:01:14.541
Many papers today, since open
sourcing these large language models,

21
00:01:14.908 --> 00:01:17.911
decided to tackle this specific challenge
on how to deal

22
00:01:17.911 --> 00:01:20.914
with outlier features for large language
models.

23
00:01:20.914 --> 00:01:25.914
Again, outlier features simply means
hidden states with large magnitude.

24
00:01:26.352 --> 00:01:30.890
So there are some interesting papers,
such as Int8, SmoothQuant,

25
00:01:32.792 --> 00:01:33.993
AWQ, and I wanted to give

26
00:01:33.993 --> 00:01:37.664
a brief explanation of each paper
to just give you some insights of

27
00:01:37.797 --> 00:01:41.101
what could be the potential solutions
to address this specific issue.

28
00:01:43.903 --> 00:01:44.771
So LLM.int8

29
00:01:44.771 --> 00:01:48.708
proposes to decompose
the underlying

30
00:01:48.741 --> 00:01:52.979
matrix multiplication
of the linear layers in two stages.

31
00:01:53.847 --> 00:01:56.649
So if you consider the input hidden states

32
00:01:56.649 --> 00:01:59.652
that you can see in the big matrix here,

33
00:02:00.019 --> 00:02:04.591
it is possible
to decompose the matmul in two parts.

34
00:02:04.591 --> 00:02:09.062
So the outlier part,
all the hidden states that are greater

35
00:02:09.062 --> 00:02:12.799
than certain threshold
and the non outlier part.

36
00:02:13.399 --> 00:02:14.834
The idea is very simple.

37
00:02:14.834 --> 00:02:17.837
So you decompose the input into

38
00:02:18.872 --> 00:02:20.507
perform the non outlier

39
00:02:20.507 --> 00:02:23.543
part matrix multiplication in Int8.

40
00:02:23.910 --> 00:02:28.882
So you quantize you do
the matmul in eight-bit and then you do

41
00:02:28.915 --> 00:02:33.915
dequantize using the scales so that you get
the final results in the input datatype.

42
00:02:34.888 --> 00:02:38.158
And the second part, you do it
classically,

43
00:02:38.691 --> 00:02:41.828
with the original dtype
of the hidden state.

44
00:02:41.861 --> 00:02:43.563
So usually in half precision.

45
00:02:43.563 --> 00:02:45.265
And then you combine both results.

46
00:02:45.265 --> 00:02:47.400
So this way
it has been proven that you can,

47
00:02:48.334 --> 00:02:49.302
retain the full

48
00:02:49.302 --> 00:02:52.705
performance of the model
without any performance degradation.

49
00:02:53.306 --> 00:02:56.442
Another very interesting approach
is called SmoothQuant.

50
00:02:56.509 --> 00:03:01.509
SmoothQuant specifically applies
to A8W8 schemes.

51
00:03:01.814 --> 00:03:05.552
Meaning
we also want to quantize the activations.

52
00:03:05.885 --> 00:03:10.190
So meaning both the activation and
the weights are in eight bit precision.

53
00:03:10.590 --> 00:03:15.395
So the paper also tackles this issue of
outlier features in large language models.

54
00:03:15.895 --> 00:03:17.830
And they proposed to mitigate that

55
00:03:17.830 --> 00:03:21.434
by smoothening
both the activation and the weights.

56
00:03:21.968 --> 00:03:26.072
Given a factor that you determine
based on the input activation

57
00:03:26.439 --> 00:03:31.411
to migrate the quantization difficulty
in both during

58
00:03:31.411 --> 00:03:35.448
the quantization of the activations,
but also quantization of the weights.

59
00:03:35.748 --> 00:03:39.752
So that way you transfer the quantization
difficulty or all over to the weights

60
00:03:40.353 --> 00:03:43.890
equally to the weights
and to the weights and the activation.

61
00:03:44.424 --> 00:03:47.927
And that way you can also retain
the full capabilities of the model.

62
00:03:48.228 --> 00:03:51.231
A more recent paper called AWQ,

63
00:03:51.564 --> 00:03:54.567
also treats
the outlier feature in a special way.

64
00:03:54.567 --> 00:03:58.738
So the paper, which came out also from
the same lab as the SmoothQuant paper,

65
00:03:59.005 --> 00:04:01.841
proposes to first iterate over a dataset

66
00:04:01.841 --> 00:04:05.612
that we are going to call
a calibration dataset

67
00:04:05.945 --> 00:04:09.182
to get detailed idea of which channel

68
00:04:10.183 --> 00:04:11.884
in the input weights

69
00:04:11.884 --> 00:04:16.589
could be responsible of generating
outlier features called salient weights.

70
00:04:17.123 --> 00:04:20.126
And the idea is,
to use that information

71
00:04:20.326 --> 00:04:23.896
to scale the model weights
before quantization,

72
00:04:24.230 --> 00:04:28.468
and also use that scale during inference
to rescale the input as well.

73
00:04:28.701 --> 00:04:30.470
So these are just a few of them.

74
00:04:30.470 --> 00:04:33.973
There are numerous other papers
that specifically address

75
00:04:34.107 --> 00:04:37.910
this issue for an effective and efficient
large language model

76
00:04:37.910 --> 00:04:38.678
quantization.

77
00:04:38.678 --> 00:04:42.782
So here is a non-exhaustive list
of those quantization techniques.

78
00:04:42.782 --> 00:04:45.718
But perhaps you can find much more
at the time we speak.

79
00:04:45.718 --> 00:04:49.689
So yeah, if you are curious about this,
I invite you to read these papers

80
00:04:49.689 --> 00:04:50.890
in detail.

81
00:04:50.890 --> 00:04:54.327
And, you know, just dive into them
and try to understand these papers.

82
00:04:54.661 --> 00:04:57.497
These are one of the challenges
when it comes to quantizing

83
00:04:57.497 --> 00:05:00.566
large language models,
because the models are quite large.

84
00:05:00.733 --> 00:05:03.336
You can get some surprising behavior.

85
00:05:03.336 --> 00:05:04.904
There are also other challenges.

86
00:05:04.904 --> 00:05:07.907
So it seems the Quantization,
Aware Training field

87
00:05:08.107 --> 00:05:11.110
seems to be a little bit
maybe underexplored today.

88
00:05:11.277 --> 00:05:14.113
So training models in low bit

89
00:05:14.113 --> 00:05:17.116
could be also
an interesting topic to dive into.

90
00:05:17.417 --> 00:05:20.920
there is also this challenge on limited
hardware support.

91
00:05:20.920 --> 00:05:25.920
So right now for this course
we only focused on W8A16 scheme,

92
00:05:26.626 --> 00:05:31.197
meaning the weights are in eight bit
but the activations are in 16 bits.

93
00:05:31.230 --> 00:05:34.200
But for a more efficient
quantization scheme, you may

94
00:05:34.200 --> 00:05:39.200
be also interested in other schemes
such as W8A8 as well.

95
00:05:39.739 --> 00:05:44.177
But not all hardwares
do support eight bit operations.

96
00:05:44.477 --> 00:05:47.313
There is also this challenge around
calibration dataset.

97
00:05:47.313 --> 00:05:51.017
So for some quantization
methods, you need to have

98
00:05:51.351 --> 00:05:54.354
a calibration dataset
to perform some sort of,

99
00:05:54.854 --> 00:05:57.990
model pre-processing
to make the quantization model better.

100
00:05:58.191 --> 00:06:01.194
And also in terms of distribution
packing and unpacking.

101
00:06:01.294 --> 00:06:03.563
So yeah, if you are really interested
about this topic,

102
00:06:03.563 --> 00:06:06.866
I invite you to do some further
reading through

103
00:06:06.866 --> 00:06:10.536
for example,
the state of the art quantization papers.

104
00:06:10.737 --> 00:06:13.940
There is also a lab called MIT Han lab,

105
00:06:14.273 --> 00:06:18.010
which made some of these, state
of the art quantization papers.

106
00:06:18.511 --> 00:06:23.483
So they have also good resources on
which you can learn more about this topic.

107
00:06:23.516 --> 00:06:24.884
You can also check out the

108
00:06:24.884 --> 00:06:28.221
Hugging Face Transformers
quantization documentation and blog post.

109
00:06:28.254 --> 00:06:32.825
You can also,
have a look at the llama.cpp repository

110
00:06:32.825 --> 00:06:37.697
discussions where you can find really
some insightful experiments and talk.

111
00:06:37.730 --> 00:06:39.031
You can also check out Reddit.

112
00:06:39.031 --> 00:06:41.534
So there is a subreddit called r/LocalLlama

113
00:06:41.534 --> 00:06:44.804
where they share a lot of cool insights
about quantization

114
00:06:45.004 --> 00:06:48.374
and you can also you can also learn more
about the new method that come up

115
00:06:48.975 --> 00:06:49.809
and so on.

116
00:06:49.809 --> 00:06:52.812
And then of course,
probably missing many more resources.

117
00:06:53.212 --> 00:06:55.581
but yeah,
these are the ones that, that I know.

118
00:06:55.581 --> 00:06:57.517
So that's it for this lesson.

119
00:06:57.517 --> 00:07:02.388
So I hope you learned a lot, through this
course and that you can use,

120
00:07:03.389 --> 00:07:06.392
the things that we have showed, to you,

121
00:07:06.426 --> 00:07:09.595
for your work or for your projects
and that all of this

122
00:07:09.595 --> 00:07:13.166
could give you some ideas of cool things
that you can do around you.

123
00:07:13.599 --> 00:07:16.869
So, yeah,
we're going to move on to the next video.

124
00:07:16.903 --> 00:07:17.236
Yeah.

125
00:07:17.236 --> 00:07:19.439
We're we'll say thank you for,

126
00:07:19.439 --> 00:07:22.442
going through this course
and suggest potential next steps.

127
00:07:22.942 --> 00:07:23.509
See you there.