Today I was able to run a nice comparison between the Prompt API in Edge (SLM Phi-4 Mini) and in Chrome (SLM Gemini Nano) — same system prompt, same settings:
- Edge makes noticeably more grammar, expression, and word choice mistakes than Chrome, at least in German.
- Edge doesn’t follow rules very well, even when they’re explicitly defined in the system prompt — and even less so when they appear as initial chat prompts. It tends to rely more on its own training data.
- Chrome, on the other hand, adheres closely to instructions and produces far fewer language errors in German.
- Chrome is faster to initialize — the session is created more quickly.
- Edge can hold more tokens in context, but that doesn’t help much if it largely ignores the information. This could be an advantage when the context contains lots of new data unfamiliar to the model, but in my use cases (product advisor systems), it didn’t make a difference.
Overall, Gemini Nano currently feels much more mature and useful than Phi-4 Mini. That might change later — both are still experimental — but for now, only Gemini Nano is truly practical.