Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -103,7 +103,7 @@ elif tokenizer == "Qwen2.5 Tokenizer":
|
|
| 103 |
with st.expander("About Qwen2.5 Tokenizer"):
|
| 104 |
st.write('')
|
| 105 |
ids = qwen_tokenizer.encode(txt)
|
| 106 |
-
split_tokens = qwen_tokenizer.
|
| 107 |
st.write(stream_data)
|
| 108 |
if token_id == True:
|
| 109 |
color = itertools.cycle(colors)
|
|
@@ -113,7 +113,7 @@ elif tokenizer == "RuAdapt Tokenizer":
|
|
| 113 |
with st.expander("About RuAdapt Tokenizer"):
|
| 114 |
st.write('')
|
| 115 |
ids = ruadapt_tokenizer.encode(txt)
|
| 116 |
-
split_tokens = ruadapt_tokenizer.
|
| 117 |
st.write(stream_data)
|
| 118 |
if token_id == True:
|
| 119 |
color = itertools.cycle(colors)
|
|
|
|
| 103 |
with st.expander("About Qwen2.5 Tokenizer"):
|
| 104 |
st.write('')
|
| 105 |
ids = qwen_tokenizer.encode(txt)
|
| 106 |
+
split_tokens = [qwen_tokenizer.decode([t]) for t in ids]
|
| 107 |
st.write(stream_data)
|
| 108 |
if token_id == True:
|
| 109 |
color = itertools.cycle(colors)
|
|
|
|
| 113 |
with st.expander("About RuAdapt Tokenizer"):
|
| 114 |
st.write('')
|
| 115 |
ids = ruadapt_tokenizer.encode(txt)
|
| 116 |
+
split_tokens = [ruadapt_tokenizer.decode([t]) for t in ids]
|
| 117 |
st.write(stream_data)
|
| 118 |
if token_id == True:
|
| 119 |
color = itertools.cycle(colors)
|