Update index.html
Browse files- index.html +11 -6
index.html
CHANGED
@@ -183,11 +183,18 @@
|
|
183 |
const [processor, model] = await SmolVLM.getInstance(modelId, dtypeSettings, device);
|
184 |
const text = processor.apply_chat_template(messages, { add_generation_prompt: true });
|
185 |
const inputs = await processor(text, [image], { do_image_splitting: doImageSplitting });
|
186 |
-
|
187 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
188 |
const streamer = new TextStreamer(processor.tokenizer, {
|
189 |
skip_prompt: true,
|
190 |
skip_special_tokens: true,
|
|
|
191 |
});
|
192 |
await model.generate({
|
193 |
...inputs,
|
@@ -195,10 +202,8 @@
|
|
195 |
min_new_tokens: maxTokens,
|
196 |
streamer,
|
197 |
});
|
198 |
-
const
|
199 |
-
|
200 |
-
|
201 |
-
const tps = maxTokens / (elapsed / 1000); // accurate TPS
|
202 |
|
203 |
totalTime += elapsed;
|
204 |
totalTps += tps;
|
|
|
183 |
const [processor, model] = await SmolVLM.getInstance(modelId, dtypeSettings, device);
|
184 |
const text = processor.apply_chat_template(messages, { add_generation_prompt: true });
|
185 |
const inputs = await processor(text, [image], { do_image_splitting: doImageSplitting });
|
186 |
+
|
187 |
+
let numTokens = 0;
|
188 |
+
let startTime;
|
189 |
+
let tps = 0;
|
190 |
+
const token_callback_function = () => {
|
191 |
+
startTime = startTime || performance.now();
|
192 |
+
tps = (numTokens++ / (performance.now() - startTime)) * 1000;
|
193 |
+
};
|
194 |
const streamer = new TextStreamer(processor.tokenizer, {
|
195 |
skip_prompt: true,
|
196 |
skip_special_tokens: true,
|
197 |
+
token_callback_function,
|
198 |
});
|
199 |
await model.generate({
|
200 |
...inputs,
|
|
|
202 |
min_new_tokens: maxTokens,
|
203 |
streamer,
|
204 |
});
|
205 |
+
const elapsed = performance.now() - start;
|
206 |
+
|
|
|
|
|
207 |
|
208 |
totalTime += elapsed;
|
209 |
totalTps += tps;
|