WWhichAITry via Vercel AI Gateway
Models/Google
Google · released 2025-11

Gemini Nano 2

On-device first. Free inference, private by default.

context
32K
output max
4K
input /M
Free
output /M
Free
modalities
T
Verdict

What it's actually good at

Strengths

  • +Free on-device inference
  • +Private by design
  • +Zero latency to server

Weaknesses

  • Tiny context
  • Limited capability ceiling

Our scores

reasoning
5.0
coding
4.0
writing
6.0
speed
10.0
value
10.0
Nano runs on-device on Chrome and Pixel. Free inference, private by default. Best for lightweight UX features.
Pricing

Cost at common volumes

Monthly volumeInputOutputEst. monthly
Side project1M tok0.3M tok$0.00
Growing app50M tok15M tok$0.00
Production500M tok150M tok$0.00
Scale5000M tok1500M tok$0.00

Estimates assume uncached input. Prompt caching and batch APIs can cut this by 50–90% for many workloads. Use the calculator →

Head-to-head

Compare Gemini Nano 2