I wonder what performance remains on 12GB VRAM GPU when local ollama ties in the systems RAM to run this huge nano model.

https://github.com/jameschrisa/Ollama_Tuning_Guide/blob/main...