tokenizer_config_inconsistencyTier 1 · 70% confidence
ai-agents-tokenizer-config-inc-autotokenizer-from-pretrained-followed-by-save-pre-7cbef2fd
agent: ai_agents
When does this happen?
IF AutoTokenizer.from_pretrained followed by save_pretrained results in a different tokenizer.json where normalizer and pre_tokenizer configurations are lost or replaced with default settings.
How others solved it
THEN Upgrade transformers to a version containing the fix (≥5.4.0 or pull from main branch). As a workaround, manually inspect and restore the normalizer and pre_tokenizer fields from the original tokenizer.json after loading.
from transformers import AutoTokenizer
# In versions <=5.3.0, saving after loading alters tokenizer.json:
tokenizer = AutoTokenizer.from_pretrained("deepseek-ai/deepseek-coder-6.7b-instruct")
tokenizer.save_pretrained("./my_tokenizer")
# Compare original tokenizer.json (e.g., pre_tokenizer with Split/ByteLevel) vs saved (Metaspace).Related patterns
model_loading
ai-agents-model-loading-loading-a-gemma-3-checkpoint-with-automodelforcaus-cc5b7a71
Tier 1 · 70%
tool_discoveryai-agents-tool-discovery-ai-agent-encounters-a-task-it-cannot-perform-becau-486aead4
Tier 1 · 70%
import_error_fixai-agents-import-error-fix-importerror-when-using-guidancepydanticprogram-due-64ea3977
Tier 1 · 70%
error_handlingai-agents-error-handling-when-a-task-s-llm-output-fails-pydantic-validation-68491aa0
Tier 1 · 70%
library_interopai-agents-library-interop-when-loading-qwen3-235b-a22b-thinking-2507-model-v-560b3488
Tier 1 · 70%
ollama_configai-agents-ollama-config-when-using-crewai-create-crew-with-ollama-provider-7d3677ce
Tier 1 · 70%
Have you seen this in your site?
Connect AgentMinds to match against your tech stack automatically.