Skip to content

Commit 9ea2c57

Browse files
committed
simplify
1 parent 643f800 commit 9ea2c57

File tree

3 files changed

+19
-34
lines changed

3 files changed

+19
-34
lines changed

.gitignore

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -144,6 +144,7 @@ ch0?/0?_user_interface/.files
144144
*.lock
145145

146146
# Temporary and OS-related files
147+
chainlit.md
147148
Untitled.ipynb
148149
.DS_Store
149150

ch05/11_qwen3/qwen3-chat-interface/qwen3-chat-interface-multiturn.py

Lines changed: 9 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -143,23 +143,15 @@ async def main(message: chainlit.Message):
143143
await out_msg.send()
144144

145145
# 3) Stream generation
146-
with torch.no_grad():
147-
for tok in generate_text_simple_stream(
148-
model=MODEL,
149-
token_ids=input_ids_tensor,
150-
max_new_tokens=MAX_NEW_TOKENS,
151-
# eos_token_id=TOKENIZER.eos_token_id
152-
):
153-
# Normalize token to int
154-
if torch.is_tensor(tok):
155-
tok = int(tok.view(-1)[0].item())
156-
else:
157-
tok = int(tok)
158-
159-
piece = TOKENIZER.decode([tok])
160-
if piece in ("<|endoftext|>", "<|im_end|>"):
161-
break
162-
await out_msg.stream_token(piece)
146+
for tok in generate_text_simple_stream(
147+
model=MODEL,
148+
token_ids=input_ids_tensor,
149+
max_new_tokens=MAX_NEW_TOKENS,
150+
eos_token_id=TOKENIZER.eos_token_id
151+
):
152+
token_id = tok.squeeze(0)
153+
piece = TOKENIZER.decode(token_id.tolist())
154+
await out_msg.stream_token(piece)
163155

164156
# 4) Finalize the streamed message
165157
await out_msg.update()

ch05/11_qwen3/qwen3-chat-interface/qwen3-chat-interface.py

Lines changed: 9 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -123,23 +123,15 @@ async def main(message: chainlit.Message):
123123
await out_msg.send()
124124

125125
# 3) Stream generation
126-
with torch.no_grad():
127-
for tok in generate_text_simple_stream(
128-
model=MODEL,
129-
token_ids=input_ids_tensor,
130-
max_new_tokens=MAX_NEW_TOKENS,
131-
):
132-
# Normalize token to int
133-
if torch.is_tensor(tok):
134-
tok = int(tok.view(-1)[0].item())
135-
else:
136-
tok = int(tok)
137-
138-
piece = TOKENIZER.decode([tok])
139-
if piece in ("<|endoftext|>", "<|im_end|>"):
140-
break
141-
142-
await out_msg.stream_token(piece)
126+
for tok in generate_text_simple_stream(
127+
model=MODEL,
128+
token_ids=input_ids_tensor,
129+
max_new_tokens=MAX_NEW_TOKENS,
130+
eos_token_id=TOKENIZER.eos_token_id
131+
):
132+
token_id = tok.squeeze(0)
133+
piece = TOKENIZER.decode(token_id.tolist())
134+
await out_msg.stream_token(piece)
143135

144136
# 4) Finalize the streamed message
145137
await out_msg.update()

0 commit comments

Comments
 (0)