File tree Expand file tree Collapse file tree 3 files changed +19
-34
lines changed
ch05/11_qwen3/qwen3-chat-interface Expand file tree Collapse file tree 3 files changed +19
-34
lines changed Original file line number Diff line number Diff line change @@ -144,6 +144,7 @@ ch0?/0?_user_interface/.files
144
144
* .lock
145
145
146
146
# Temporary and OS-related files
147
+ chainlit.md
147
148
Untitled.ipynb
148
149
.DS_Store
149
150
Original file line number Diff line number Diff line change @@ -143,23 +143,15 @@ async def main(message: chainlit.Message):
143
143
await out_msg .send ()
144
144
145
145
# 3) Stream generation
146
- with torch .no_grad ():
147
- for tok in generate_text_simple_stream (
148
- model = MODEL ,
149
- token_ids = input_ids_tensor ,
150
- max_new_tokens = MAX_NEW_TOKENS ,
151
- # eos_token_id=TOKENIZER.eos_token_id
152
- ):
153
- # Normalize token to int
154
- if torch .is_tensor (tok ):
155
- tok = int (tok .view (- 1 )[0 ].item ())
156
- else :
157
- tok = int (tok )
158
-
159
- piece = TOKENIZER .decode ([tok ])
160
- if piece in ("<|endoftext|>" , "<|im_end|>" ):
161
- break
162
- await out_msg .stream_token (piece )
146
+ for tok in generate_text_simple_stream (
147
+ model = MODEL ,
148
+ token_ids = input_ids_tensor ,
149
+ max_new_tokens = MAX_NEW_TOKENS ,
150
+ eos_token_id = TOKENIZER .eos_token_id
151
+ ):
152
+ token_id = tok .squeeze (0 )
153
+ piece = TOKENIZER .decode (token_id .tolist ())
154
+ await out_msg .stream_token (piece )
163
155
164
156
# 4) Finalize the streamed message
165
157
await out_msg .update ()
Original file line number Diff line number Diff line change @@ -123,23 +123,15 @@ async def main(message: chainlit.Message):
123
123
await out_msg .send ()
124
124
125
125
# 3) Stream generation
126
- with torch .no_grad ():
127
- for tok in generate_text_simple_stream (
128
- model = MODEL ,
129
- token_ids = input_ids_tensor ,
130
- max_new_tokens = MAX_NEW_TOKENS ,
131
- ):
132
- # Normalize token to int
133
- if torch .is_tensor (tok ):
134
- tok = int (tok .view (- 1 )[0 ].item ())
135
- else :
136
- tok = int (tok )
137
-
138
- piece = TOKENIZER .decode ([tok ])
139
- if piece in ("<|endoftext|>" , "<|im_end|>" ):
140
- break
141
-
142
- await out_msg .stream_token (piece )
126
+ for tok in generate_text_simple_stream (
127
+ model = MODEL ,
128
+ token_ids = input_ids_tensor ,
129
+ max_new_tokens = MAX_NEW_TOKENS ,
130
+ eos_token_id = TOKENIZER .eos_token_id
131
+ ):
132
+ token_id = tok .squeeze (0 )
133
+ piece = TOKENIZER .decode (token_id .tolist ())
134
+ await out_msg .stream_token (piece )
143
135
144
136
# 4) Finalize the streamed message
145
137
await out_msg .update ()
You can’t perform that action at this time.
0 commit comments