aimlbd / HuggingFace.Quantization_in_Depth /Transcript /7_Per Group Quantization_HFQD.txt
swsuws's picture
Upload 200 files
0c08f5a verified
raw
history blame
9.78 kB
WEBVTT
X-TIMESTAMP-MAP=LOCAL:00:00:00.000,MPEGTS:144533
1
00:00:02.002 --> 00:00:03.003
Now, let's go
2
00:00:03.003 --> 00:00:05.972
even smaller and do per group
quantization.
3
00:00:05.972 --> 00:00:10.777
In per group quantization we perform
quantization on groups of n elements.
4
00:00:11.144 --> 00:00:13.813
Common values for n are 32,
5
00:00:13.813 --> 00:00:16.883
64, or 128. Per group
6
00:00:16.883 --> 00:00:19.619
quantization can require a lot of memory.
7
00:00:19.619 --> 00:00:20.086
Let's say,
8
00:00:20.086 --> 00:00:25.086
we want to quantize a tensor in four-bit,
and we choose a group size equal to 32.
9
00:00:25.325 --> 00:00:26.459
We use symmetric mode.
10
00:00:26.459 --> 00:00:29.062
That means that the zero point
is equal to zero,
11
00:00:29.062 --> 00:00:32.265
and we store the scales
in floating point 16.
12
00:00:32.599 --> 00:00:37.504
It means that we are actually quantizing
the tensor in 4.5 bits.
13
00:00:37.871 --> 00:00:42.709
Since we have four bits, since each
element is stored using four bit
14
00:00:43.343 --> 00:00:47.414
and we have 16 divided by 32 bit.
15
00:00:47.747 --> 00:00:51.985
Since we need to store a scale in 16 bits
16
00:00:51.985 --> 00:00:56.823
for every 32 elements for each element,
you store it in four bit,
17
00:00:56.923 --> 00:01:01.923
but you also have quantization parameters
and you need to store once
18
00:01:02.462 --> 00:01:07.033
a scale in 16 bits,
so 16 bits every 32 elements.
19
00:01:07.167 --> 00:01:09.069
Now let's jump to the code.
20
00:01:09.069 --> 00:01:12.305
For simplicity,
we will restrict ourselves to the case
21
00:01:12.305 --> 00:01:16.709
where the tensor is of dimension two
and we will be using the symmetric mode.
22
00:01:16.943 --> 00:01:17.210
You don't
23
00:01:17.210 --> 00:01:20.680
need to pay attention to this code
since we will be coding in the notebook.
24
00:01:20.980 --> 00:01:22.615
Now let's code it.
25
00:01:22.615 --> 00:01:25.518
So we define the following function.
26
00:01:25.518 --> 00:01:27.353
Linear q symmetric per group.
27
00:01:28.955 --> 00:01:31.958
This will take as argument the tensor,
28
00:01:33.126 --> 00:01:36.129
the group size and the d-type.
29
00:01:37.797 --> 00:01:41.000
We set the default value torch.int8.
30
00:01:42.502 --> 00:01:46.239
First,
we need to get the shape of the tensor.
31
00:01:50.276 --> 00:01:53.413
Then, another restriction for this function
32
00:01:53.613 --> 00:01:57.750
is that we will be performing
quantization on the rows.
33
00:01:57.851 --> 00:02:02.851
This is why we also need to make sure
that each row is divisible by group size.
34
00:02:02.922 --> 00:02:05.892
To confirm that,
we will just use assertion
35
00:02:05.892 --> 00:02:10.892
so that the shape of the tensor along the
rows is indeed divisible by group size.
36
00:02:13.366 --> 00:02:17.637
Then, as I said,
we will be restricting ourselves
37
00:02:17.637 --> 00:02:20.640
to tensors of dimension two.
38
00:02:22.509 --> 00:02:26.646
Now all we need to do
is to reshape the tensor
39
00:02:26.980 --> 00:02:31.084
so that we end up with rows of group
size elements.
40
00:02:31.784 --> 00:02:36.122
To do that, we will use the view function
that we learned about.
41
00:02:40.527 --> 00:02:42.862
So as you can see, what we do here
42
00:02:42.862 --> 00:02:46.499
is to make sure that each row contains
group size elements.
43
00:02:46.799 --> 00:02:49.702
And we put the minus one here
so that it infers
44
00:02:49.702 --> 00:02:53.273
automatically the right dimension
to have in the first dimension.
45
00:02:53.540 --> 00:02:57.310
And now if you look at the tensor
we has the setup
46
00:02:57.577 --> 00:03:00.413
for performing positional quantization.
47
00:03:00.413 --> 00:03:04.984
We resized this tensor
so that we have rows of group size
48
00:03:05.084 --> 00:03:09.088
so that we can use the function
that we coded previously.
49
00:03:09.522 --> 00:03:12.959
That is to say the linear q
symmetric channel quantization.
50
00:03:13.126 --> 00:03:16.129
So we have quantized tensor
51
00:03:16.763 --> 00:03:17.864
and scale
52
00:03:17.864 --> 00:03:21.768
which is equal to linear
q symmetric per channel function.
53
00:03:22.669 --> 00:03:26.506
And we need to put the tensor
the right dimension.
54
00:03:26.906 --> 00:03:29.976
So along the rows and the d-type.
55
00:03:33.179 --> 00:03:35.348
After quantizing the tensor
56
00:03:35.348 --> 00:03:38.551
we still need to reshape it
to its original shape.
57
00:03:38.851 --> 00:03:41.688
So we will use the shape
that we stored before.
58
00:03:41.688 --> 00:03:43.022
Here the d shape.
59
00:03:46.292 --> 00:03:47.594
To reshape the tensor,
60
00:03:47.594 --> 00:03:51.130
we use the view
and we just pass this shape.
61
00:03:51.564 --> 00:03:54.567
Then we can return the quantized tensor
and the scale.
62
00:03:55.902 --> 00:03:57.570
Now that we have coded
63
00:03:57.570 --> 00:04:02.342
the per group quantization, now let's code
the linear quantization
64
00:04:02.575 --> 00:04:07.380
for the quantization
in order to verify our results.
65
00:04:07.947 --> 00:04:10.950
So we need to define
66
00:04:10.950 --> 00:04:11.951
this function.
67
00:04:11.951 --> 00:04:15.021
In that function
we need the quantized tensor
68
00:04:16.022 --> 00:04:18.658
to scale.
69
00:04:18.658 --> 00:04:21.661
But we also need the group size.
70
00:04:25.098 --> 00:04:29.068
Then we need to get the shape
of the quantized tensor.
71
00:04:29.102 --> 00:04:32.105
That will be useful.
72
00:04:33.106 --> 00:04:35.842
Then we need to reshape
73
00:04:35.842 --> 00:04:39.479
the quantized tensor
so that we have rows that contain
74
00:04:39.479 --> 00:04:42.482
only group size elements.
75
00:04:42.548 --> 00:04:46.986
To do that,
we put in the view methods minus
76
00:04:46.986 --> 00:04:51.190
one for the first value and group size
for the second one.
77
00:04:52.191 --> 00:04:56.062
Then we can reuse the linear
78
00:04:56.162 --> 00:05:00.366
dequantization methods
we coded before to dequantize the tensor.
79
00:05:00.600 --> 00:05:05.600
We need to pass the quantized tensor,
the scale and decimal point.
80
00:05:06.072 --> 00:05:08.641
But since we are
doing symmetric quantization,
81
00:05:09.942 --> 00:05:11.878
the zero point is equal to zero.
82
00:05:11.878 --> 00:05:16.382
Then all we need to do is to reshape
the dequantized tensor
83
00:05:16.749 --> 00:05:21.554
with the shape of the original tensor,
and the shape is stored in q shape.
84
00:05:24.223 --> 00:05:27.226
Then we return the dequantized tensor.
85
00:05:27.994 --> 00:05:30.997
Now let's test our implementation.
86
00:05:30.997 --> 00:05:33.933
We will test a random tensor of size six
87
00:05:33.933 --> 00:05:36.903
by six and
88
00:05:37.937 --> 00:05:40.840
let's set group size to be equal to three.
89
00:05:40.840 --> 00:05:45.478
So, we will get the quantized tensor
and the scale
90
00:05:46.546 --> 00:05:50.450
using the linear q symmetric group function.
91
00:05:50.850 --> 00:05:53.553
And we need to pass the test
92
00:05:53.553 --> 00:05:56.556
tensor as well as the group size.
93
00:05:56.756 --> 00:06:00.293
Then to verify our results
we also need to
94
00:06:00.293 --> 00:06:03.296
dequantize the tensor using
95
00:06:03.296 --> 00:06:07.500
the linear dequantization function
where we need to pass
96
00:06:08.101 --> 00:06:10.737
the quantized tensor,
97
00:06:10.737 --> 00:06:13.306
the scale, and the group size.
98
00:06:13.306 --> 00:06:16.743
Finally, to have the summary
of the quantization process,
99
00:06:17.377 --> 00:06:20.380
we just need to pass inside the plot
quantization error.
100
00:06:20.813 --> 00:06:23.216
The following arguments.
101
00:06:23.216 --> 00:06:26.419
So test tensor,
quantized tensor and dequantized tensor.
102
00:06:26.853 --> 00:06:31.224
And as you can see,
if you look at the quantized tensor
103
00:06:31.557 --> 00:06:34.761
you will see that
every three elements in each row
104
00:06:35.161 --> 00:06:38.364
you will have the maximum value 127.
105
00:06:38.965 --> 00:06:42.135
It shows that we indeed managed
to quantize
106
00:06:42.435 --> 00:06:45.738
each three elements in this matrix
along the rows.
107
00:06:45.905 --> 00:06:49.442
So three elements
here, three here, and so on.
108
00:06:49.742 --> 00:06:51.411
And you have the quantized tensor.
109
00:06:51.411 --> 00:06:52.979
As you can see on the right.
110
00:06:52.979 --> 00:06:57.979
And you can see also that the quantization
error tensor is very, very low
111
00:06:58.518 --> 00:07:01.521
and that the dequantized tensor is
112
00:07:01.988 --> 00:07:04.991
practically the same
as the original tensor.
113
00:07:05.057 --> 00:07:09.262
Let's also print the dequantization error
using the dequantization error function.
114
00:07:09.529 --> 00:07:12.632
And we just need to pass the test tensor
and the quantized tensor.
115
00:07:13.933 --> 00:07:16.903
And indeed we have a very very low
quantization error.
116
00:07:16.903 --> 00:07:20.573
Now is a good time to pause the video
and try a couple of things.
117
00:07:20.606 --> 00:07:23.042
You can try to change the test tensors.
118
00:07:23.042 --> 00:07:25.311
Or you can also change the group size.
119
00:07:25.311 --> 00:07:30.311
And to see what is the effect of the
group size on the dequantization process.