Update index.html
Browse files- index.html +11 -6
index.html
CHANGED
|
@@ -183,11 +183,18 @@
|
|
| 183 |
const [processor, model] = await SmolVLM.getInstance(modelId, dtypeSettings, device);
|
| 184 |
const text = processor.apply_chat_template(messages, { add_generation_prompt: true });
|
| 185 |
const inputs = await processor(text, [image], { do_image_splitting: doImageSplitting });
|
| 186 |
-
|
| 187 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 188 |
const streamer = new TextStreamer(processor.tokenizer, {
|
| 189 |
skip_prompt: true,
|
| 190 |
skip_special_tokens: true,
|
|
|
|
| 191 |
});
|
| 192 |
await model.generate({
|
| 193 |
...inputs,
|
|
@@ -195,10 +202,8 @@
|
|
| 195 |
min_new_tokens: maxTokens,
|
| 196 |
streamer,
|
| 197 |
});
|
| 198 |
-
const
|
| 199 |
-
|
| 200 |
-
|
| 201 |
-
const tps = maxTokens / (elapsed / 1000); // accurate TPS
|
| 202 |
|
| 203 |
totalTime += elapsed;
|
| 204 |
totalTps += tps;
|
|
|
|
| 183 |
const [processor, model] = await SmolVLM.getInstance(modelId, dtypeSettings, device);
|
| 184 |
const text = processor.apply_chat_template(messages, { add_generation_prompt: true });
|
| 185 |
const inputs = await processor(text, [image], { do_image_splitting: doImageSplitting });
|
| 186 |
+
|
| 187 |
+
let numTokens = 0;
|
| 188 |
+
let startTime;
|
| 189 |
+
let tps = 0;
|
| 190 |
+
const token_callback_function = () => {
|
| 191 |
+
startTime = startTime || performance.now();
|
| 192 |
+
tps = (numTokens++ / (performance.now() - startTime)) * 1000;
|
| 193 |
+
};
|
| 194 |
const streamer = new TextStreamer(processor.tokenizer, {
|
| 195 |
skip_prompt: true,
|
| 196 |
skip_special_tokens: true,
|
| 197 |
+
token_callback_function,
|
| 198 |
});
|
| 199 |
await model.generate({
|
| 200 |
...inputs,
|
|
|
|
| 202 |
min_new_tokens: maxTokens,
|
| 203 |
streamer,
|
| 204 |
});
|
| 205 |
+
const elapsed = performance.now() - start;
|
| 206 |
+
|
|
|
|
|
|
|
| 207 |
|
| 208 |
totalTime += elapsed;
|
| 209 |
totalTps += tps;
|