What is your worst-case scenario here?

Something like a pop-sci article along the lines of "Mad scientists create racist, imperialistic AI"?

I honestly don't see publication of the weights as a relevant risk factor, because sensationalist misrepresentation is trivially possible with the given example responses alone.

I don't think such pseudo-malicious misrepresentation of scientific research can be reliably prevented anyway, and the disclaimers make your stance very clear.

On the other hand, publishing weights might lead to interesting insights from others tinkering with the models. A good example for this would be the published word prevalence data (M. Brysbaert et al @Ghent University) that led to interesting follow-ups like this: https://observablehq.com/@yurivish/words

I hope you can get the models out in some form, would be a waste not to, but congratulations on a fascinating project regardless!

It seems like if there is an obvious misuse of a tool, one has a moral imperative to restrict use of the tool.

Every tool can be misused. Hammers are as good for bashing heads as building houses. Restricting hammers would be silly and counterproductive.

Yes but if you are building an voice activated autonomous flying hammer then you either want it to be very good at differentiating heads from hammers OR you should restrict its use.

OR you respect individual liberty and agency, hold individuals responsible for their actions, instead of tools, and avoid becoming everyone's condescending nanny.

Your pre-judgement of acceptable hammer uses would rob hammer owners of responsible and justified self-defense and defense of others in situations in which there are no other options, as well as other legally and socially accepted uses which do not fit your pre-conceived ideas.