By default, Ollama uses a context window size of 2048 tokens.
I suspect the Ollama version might have wrong default settings, such as conversation delimiters. The experience of Gemma 3 in AI studio is completely different.
By default, Ollama uses a context window size of 2048 tokens.