Nono Martínez Alonso
marked this post as
complete
Nono Martínez Alonso
I published a video on how to run Google Gemma 2B & 7B on Apple Silicon with Python at https://www.youtube.com/watch?v=qFULISWcjQc.
People have said using Gemma via Ollama is much faster, and my suspicion is that they're using it via llama.cpp or gemma.cpp.
Nono Martínez Alonso
marked this post as
in progress