Failed to run local embedding model using llama.cpp #22210
Unanswered
Wei-W2025-code
asked this question in
Q&A
Replies: 1 comment
-
|
"Invalid tokens" on embeddings usually points at the tokenizer/vocab, not the request format. Likely causes: vocab mismatch if the GGUF was re-quantized without matching |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Strat Code :
My pthon code:
Service OutPut :
My Cmd Answer:
Beta Was this translation helpful? Give feedback.
All reactions