홈으로 돌아가기
Hacker News

구글, Gemma 4 오픈 모델 출시

Google releases Gemma 4 open models

1704 points 450 comments jeffmcjunkin 2026-04-03 01:10

댓글

7
danielhanchen 2026-04-03 01:16
ENGLISH (원문)
Thinking / reasoning + multimodal + tool calling. We made some quants at https://huggingface.co/collections/unsloth/gemma-4 for folks to run them - they work really well! Guide for those interested: https://unsloth.ai/docs/models/gemma-4 Also note to use temperature = 1.0, top_p = 0.95, top_k = 64 and the EOS is "<turn|>". "<|channel>thought\n" is also used for the thinking trace!
jwr 2026-04-03 01:17
ENGLISH (원문)
Really looking forward to testing and benchmarking this on my spam filtering benchmark. gemma-3-27b was a really strong model, surpassed later by gpt-oss:20b (which was also much faster). qwen models always had more variance.
a7om_com 2026-04-03 01:25
ENGLISH (원문)
Gemma models are already in our AIPI inference pricing index. Open source models like Gemma run 70.7% cheaper than proprietary equivalents at the median across the 2,614 SKUs we track. With Gemma 4 hitting third-party platforms the pricing will be worth watching closely. Full data at a7om.com.
flakiness 2026-04-03 01:25
ENGLISH (원문)
It's good they still have non-instruction-tuned models.
minimaxir 2026-04-03 01:25
ENGLISH (원문)
The benchmark comparisons to Gemma 3 27B on Hugging Face are interesting: The Gemma 4 E4B variant (https://huggingface.co/google/gemma-4-E4B-it) beats the old 27B in every benchmark at a fraction of parameters. The E2B/E4B models also support voice input, which is rare.
NitpickLawyer 2026-04-03 01:25
ENGLISH (원문)
Best thing is that this is Apache 2.0 The sizes are E2B and E4B (following gemma3n arch, with focus on mobile) and 26BA4 MoE and 31B dense. The mobile ones have audio in (so I can see some local privacy focused translation apps) and the 31B seems to be strong in agentic stuff. 26BA4 stands somewhere in between, similar VRAM footprint, but much faster inference.
babelfish 2026-04-03 01:25
ENGLISH (원문)
Wow, 30B parameters as capable as a 1T parameter model?