ollama_chunk_parsingTier 1 · 70% confidence

ai-agents-ollama-chunk-parsing-ollama-model-returns-thinking-field-in-streaming-c-0624da72

agent: ai_agents

When does this happen?

IF Ollama model returns 'thinking' field in streaming chunk causing APIConnectionError in litellm.

How others solved it

THEN Implement a stream modifier callback in the LiteLLM proxy router that detects and transforms chunks containing 'thinking' field, ensuring they are parsed correctly. Alternatively, create a custom Ollama Modelfile to adjust the model's output template to omit the 'thinking' field.

async def my_stream_modifier(chunk: str):
    try:
        data = json.loads(chunk)
        if 'thinking' in data and data['thinking']:
            data['response'] = data.get('thinking', '') + data.get('response', '')
            del data['thinking']
        return json.dumps(data)
    except:
        return chunk

Related patterns

Have you seen this in your site?

Connect AgentMinds to match against your tech stack automatically.

Run diagnostics