monurcan commited on
Commit
dc798af
·
1 Parent(s): 20b327e
Files changed (1) hide show
  1. app.py +11 -2
app.py CHANGED
@@ -81,10 +81,12 @@ def model_inference(input_dict, history):
81
  streamer = TextIteratorStreamer(
82
  processor, skip_prompt=True, skip_special_tokens=True
83
  )
84
- generation_kwargs = dict(inputs, streamer=streamer, max_new_tokens=1024)
 
 
85
  thread = Thread(target=model.generate, kwargs=generation_kwargs)
86
  thread.start()
87
- buffer = ""
88
  yield progress_bar_html("Processing...")
89
  for new_text in streamer:
90
  escaped_new_text = html.escape(new_text)
@@ -93,6 +95,13 @@ def model_inference(input_dict, history):
93
  time.sleep(0.001)
94
  yield buffer
95
 
 
 
 
 
 
 
 
96
 
97
  examples = [
98
  [
 
81
  streamer = TextIteratorStreamer(
82
  processor, skip_prompt=True, skip_special_tokens=True
83
  )
84
+ generation_kwargs = dict(inputs, streamer=streamer, max_new_tokens=2048)
85
+ # start timer just before generation begins
86
+ start_time = time.time()
87
  thread = Thread(target=model.generate, kwargs=generation_kwargs)
88
  thread.start()
89
+ buffer = "Baseline Model Response: "
90
  yield progress_bar_html("Processing...")
91
  for new_text in streamer:
92
  escaped_new_text = html.escape(new_text)
 
95
  time.sleep(0.001)
96
  yield buffer
97
 
98
+ # Ensure generation thread has finished and measure elapsed time
99
+ thread.join()
100
+ elapsed = time.time() - start_time
101
+ elapsed_text = f"\nBaseline generation time: {elapsed:.2f} s"
102
+ buffer += html.escape(elapsed_text)
103
+ yield buffer
104
+
105
 
106
  examples = [
107
  [