I wonder what performance remains on 12GB VRAM GPU when local ollama ties in the systems RAM to run this huge nano model.
https://github.com/jameschrisa/Ollama_Tuning_Guide/blob/main...
I wonder what performance remains on 12GB VRAM GPU when local ollama ties in the systems RAM to run this huge nano model.
https://github.com/jameschrisa/Ollama_Tuning_Guide/blob/main...