Files
ollama/fs
Daniel Hiltgen e823bff873 gemma4: enable flash attention (#15378)
Backport GGML kernels so we can enable flash attention for the gemma 4 model on
Metal and CUDA.
2026-04-07 08:12:36 -07:00
..
2025-02-13 16:31:21 -08:00