Google is probably still sucking up the contents of your LLM requests even with the model running locally.