andito HF Staff commited on
Commit
c06b59f
·
verified ·
1 Parent(s): 6ba7708

Update index.html

Browse files
Files changed (1) hide show
  1. index.html +11 -6
index.html CHANGED
@@ -183,11 +183,18 @@
183
  const [processor, model] = await SmolVLM.getInstance(modelId, dtypeSettings, device);
184
  const text = processor.apply_chat_template(messages, { add_generation_prompt: true });
185
  const inputs = await processor(text, [image], { do_image_splitting: doImageSplitting });
186
-
187
- const start = performance.now();
 
 
 
 
 
 
188
  const streamer = new TextStreamer(processor.tokenizer, {
189
  skip_prompt: true,
190
  skip_special_tokens: true,
 
191
  });
192
  await model.generate({
193
  ...inputs,
@@ -195,10 +202,8 @@
195
  min_new_tokens: maxTokens,
196
  streamer,
197
  });
198
- const end = performance.now();
199
- const elapsed = end - start;
200
-
201
- const tps = maxTokens / (elapsed / 1000); // accurate TPS
202
 
203
  totalTime += elapsed;
204
  totalTps += tps;
 
183
  const [processor, model] = await SmolVLM.getInstance(modelId, dtypeSettings, device);
184
  const text = processor.apply_chat_template(messages, { add_generation_prompt: true });
185
  const inputs = await processor(text, [image], { do_image_splitting: doImageSplitting });
186
+
187
+ let numTokens = 0;
188
+ let startTime;
189
+ let tps = 0;
190
+ const token_callback_function = () => {
191
+ startTime = startTime || performance.now();
192
+ tps = (numTokens++ / (performance.now() - startTime)) * 1000;
193
+ };
194
  const streamer = new TextStreamer(processor.tokenizer, {
195
  skip_prompt: true,
196
  skip_special_tokens: true,
197
+ token_callback_function,
198
  });
199
  await model.generate({
200
  ...inputs,
 
202
  min_new_tokens: maxTokens,
203
  streamer,
204
  });
205
+ const elapsed = performance.now() - start;
206
+
 
 
207
 
208
  totalTime += elapsed;
209
  totalTps += tps;