Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
13.4
TFLOPS
25
1
17
Lucas Silva
PRO
YuLexuan30
Follow
f5139's profile picture
upgraedd's profile picture
ClayFace's profile picture
7 followers
·
7 following
AI & ML interests
None yet
Recent Activity
new
activity
2 days ago
black-forest-labs/FLUX.2-dev:
Space no longer working
reacted
to
martinsu
's
post
with 🔥
2 days ago
I wasted days on a GPU node on a bug that shouldn't exist So I was fine-tuning TildeOPEN-30B and the outputs were... weird. Token ID 179 (<0x00>) kept appearing between almost every token pair. Took me a bit to figure out what was going on. Turns out I used the fast tokenizer for training, but the model was trained on the slow one. Silent failure. Well... long story short—TGI uses (forces) the fast tokenizer, no questions asked. And you'll have agile's kryptonite: silent failure. If the model was trained on slow, it's a silent disaster. I got curious and wrote a quick script to check how common this is. Ran it on 6,014 LLM HF models overnight. Roughly 10% of HF model downloads have mismatched tokenizers. Not all mismatches are catastrophic, but some are brutal — like chat template markers inflating from 1 token to 3, silently wrecking context windows and causing model act weird. This wasn't rigorous research, but the drift is real. And the worst part? 968 models(out of 500+ downloads) have both fast and slow tokenizers present, but they still produce different outputs. No missing files, no errors — just silent degradation. TGI defaults to the fast tokenizer, as does AutoTokenizer.from_pretrained(). If a fast tokenizer doesn't exist, it auto-generates one. If your model was trained on slow, you get silent degradation. Output looks fine; the model just performs worse. Sometimes really worse. You'd never know. If model was trained on fast tokenizer, its fine, but how do You know? The root cause? Either model authors run HF conversion and upload both without verifying, or users run TGI, which always forces(converts to) fast . The result of this fight with tokenizers is https://huggingface.co/martinsu/tildeopen-30b-mu-instruct It's based on TildeOPEN-30B (a solid EU HPC multilingual base). Nothing fancy—just a proper instruction fine-tune where I didn't mess up the tokenizer this time. Full article: https://github.com/martins-u/tokenmagedon
liked
a Space
13 days ago
black-forest-labs/FLUX.2-dev
View all activity
Organizations
models
12
Sort: Recently updated
YuLexuan30/yuxuanchroma-lora
Text-to-Image
•
Updated
29 days ago
•
115
YuLexuan30/tangbohuchroma-lora
Text-to-Image
•
Updated
Oct 19
•
24
YuLexuan30/chromaloonerv3-lora
Text-to-Image
•
Updated
Oct 12
•
37
YuLexuan30/desired_lora_FK_Crafty_Estate2088
Updated
Jul 5
•
11
YuLexuan30/Meng_Zhang_Lora_MenWithSwords
Text-to-Image
•
Updated
May 3
•
3
YuLexuan30/Yu_Lexuan_Archer
Text-to-Image
•
Updated
Apr 30
•
13
•
YuLexuan30/Yu_Lexuan_Flux_0.1
Text-to-Image
•
Updated
Apr 25
•
4
•
YuLexuan30/Feng_Long_Flux
Text-to-Image
•
Updated
Apr 12
•
7
•
YuLexuan30/Looner_Concept_Lora_SDXL_Experimental_Rank_2
Updated
Apr 10
YuLexuan30/Yu_Lexuan
Text-to-Image
•
Updated
Mar 19
•
11
•
•
1
View 12 models
datasets
24
Sort: Recently updated
YuLexuan30/yuxuanchroma-dataset
Viewer
•
Updated
29 days ago
•
36
•
74
YuLexuan30/heartballoonwanbig-dataset
Viewer
•
Updated
Nov 8
•
79
•
84
YuLexuan30/wanheartballoonv5-dataset
Updated
Nov 8
•
109
YuLexuan30/vadimshatilovchromav1-dataset
Viewer
•
Updated
Oct 30
•
16
•
8
YuLexuan30/tangbohuchroma-dataset
Viewer
•
Updated
Oct 19
•
37
•
17
YuLexuan30/loonerchromagemini-dataset
Viewer
•
Updated
Oct 16
•
189
•
20
YuLexuan30/looner190v2
Viewer
•
Updated
Oct 12
•
189
•
18
YuLexuan30/wanbtp81f-dataset
Viewer
•
Updated
Oct 11
•
16
•
31
YuLexuan30/yokokuramachromav1-dataset
Viewer
•
Updated
Oct 8
•
19
•
18
YuLexuan30/marquis_v1-dataset
Viewer
•
Updated
Oct 6
•
10
•
21
View 24 datasets