Tokenization Issues with inital configs provided.

#1
by Nitral-AI - opened

Reverted config files back with original bmo's, as i was seeing endless generation with the chatml modified configs. The way i swapped the embeddings out in the chatmlificaton was kind of hacky and still very experimental (token-surgery using violet_twighlight-v2 as the donor model to approximate chatml embeddings and replaced the embeddings inside of the original bmo.) [Quants will need to be redone.] @bartowski @Lewdiculous (Should still work in chatml during inference with the reverted changes, since the embedding themselves should reflect chatml approximations due to the 'surgery' and merge itself.)

Nitral-AI changed discussion status to closed
Nitral-AI changed discussion title from The tokenization problems from the original chatmlification of bmo has leaked into this. to The tokenization problems from the original chatmlification of bmo may have leaked into this.
This comment has been hidden
Nitral-AI changed discussion status to open
Nitral-AI changed discussion title from The tokenization problems from the original chatmlification of bmo may have leaked into this. to Tokenization Issues with inital configs provided.

Reverted config files back with original bmo's, as i was seeing endless generation with the chatml modified configs.

That explains it, yeah... It was way too late and I was way too sleepy already to really test much or say anything, haha. Thanks for the heads up.

Sorry for all the trouble to both of you. Ill make sure to mark the models with an experimental tag, until they are done with the full suite of testing going foreword.

All good!
v2 already finishing uploads on my end.

No problem @Nitral-AI , mine is also updated, appreciate the ping :)

Sign up or log in to comment