Interesting, would you mind sharing? (imgur allows free image uploads, quick drag and drop)
I do have a "works on my machine"* :) -- prompt "Model F keyboard", all settings disabled, on the smaller model, seems to have substantially more than no idea: https://imgur.com/a/32pV6Sp
(Google Images comparison included to show in-the-wild "Model F keyboard", which may differ from my/your expected distribution)
* my machine, being, https://playground.bfl.ai/ (I have no affiliation with BFL)
Your generated examples just look like generic modern-day mechanical keyboards, they don't have any of the Model Fs defining features.
Your Google Images search indicates the original problem of models training on junk misinformation online. If AI scrapers are downloading every photo that's associated with "Model F Keyboard" like that, the models have no idea what is an IBM Model F, or its distinguishing characteristics, and what is some other company's, and what is misidentified.
https://commons.wikimedia.org/wiki/Category:IBM_Model_F_Keyb...
Specifying "IBM Model F keyboard" and placing it in quotation marks improves the search. But the front page of the search is tip-of-the-iceberg compared to whatever the model's scrapers ingested.
Eventually you may hit trademark protections. Reproducing a brand-name keyboard may be as difficult as simulating a celebrity's likeness.
I'm not even sure what my friends look like on Facebook, so it's not clear how an AI model would reproduce a brand-name keyboard design on request.
I agree with you vehemently.
Another way of looking at it is, insistence on complete verisimilitude in an image generator is fundamentally in error.
I would argue, even undesirable. I don't want to live in a world where a 45 year old keyboard that was only out for 4 years is readily imitated in every microscopic detail.
I also find myself frustrated, and asking myself why.
First thought that jumps in: it's very clear that it is in error to say the model has no idea, modulo there's some independent run that's dramatically different from the only one offered in this thread.
Second thought: if we're doing "the image generators don't get details right", it would seem to be there a lot simpler examples than OPs, and it is better expressed that way - I assume it wasn't expressed that way because it sounds like dull conversation, but it doesn't have to be!
Third thought as to why I feel frustrated: I feel like I wasted time here - no other demos showing it's anywhere close to "no idea", its completely unclear to me whats distinctive about a IBM Model F Keyboard, and the the wikipedia images are worse than Google's AFAICT.
> if we're doing "the image generators don't get details right", it would seem to be there a lot simpler examples than OPs
There are different sorts of details though, and the distinctions are both useful and interesting to understanding the state of the art. If "man drinking coke" produces someone with 6 fingers holding a glass of water that's completely different from producing someone with 5 fingers holding a can of pepsi.
Notice that none of the images in your example got the function key placement correct. Clearly the model knows what a relatively modern keyboard is, and it even has some concept of a vaguely retro looking mechanical keyboard. However indeed I'm inclined to agree with OP that it has approximately zero idea what an "IBM model F" keyboard is. I'm not sure that's a failure of the model though - as you point out, it's an ancient and fairly obscure product.
> Eventually you may hit trademark protections. Reproducing a brand-name keyboard may be as difficult as simulating a celebrity's likeness.
Then the law is broken. Monetizing someone's likeness is an issue. Utilizing trademarked characteristics to promote your own product without permission is an issue. It's the downstream actions of the user that are the issue, not the ML model itself.
Models regurgitating copyrighted material verbatim is of course an entirely separate issue.
>> Eventually you may hit trademark protections.
> Then the law is broken.
> Utilizing trademarked characteristics to promote your own product without permission is an issue.
It sounds like you agree with parent that if your product reproduces trademark characteristics, it is utilizing trademarked characteristics. Just about at what layer you don't have responsibility. And the layer that has responsibility is the one that profits unjustly from the AI.
I'm interested if there's an argument for saying only the 2nd party user of the 1st party AI model, selling AI model output, to a 3rd party, is intuitively unfair.
I can't think of one. e.g. Disney launches some new cartoon or whatever. 1st party Openmetagoog, trains on it to make my "Video Episode Generator" product. Now, Openmetagoogs Community Pages are full of 30m video episodes made by their image generator. They didn't make them, nor do they promote them. Inuitively, Openmetagoog a competitor for manufacturing my IP, and that is also intuitively wrong. Your analysis would have us charge the users for sharing the output.
> if your product reproduces trademark characteristics, it is utilizing trademarked characteristics.
I wouldn't agree with that, no. To my mind "utilizing" generally requires intent at least in the context we're discussing here (ie moral or legal obligations). I'd remind you that the entire point of trademark is (approximately) to prevent brand confusion within the market.
> Your analysis would have us charge the users for sharing the output.
Precisely. I see it as both a matter of intent and concrete damages. Creating something (pencil, diffusion model, camera, etc) that could possibly be used in a manner that violates the law is not a problem. It is the end user violating the law that is at fault.
Imagine an online community that uses blender to create disney knockoffs and shares them publicly. Blender is not at fault and the creation of the knockoffs themselves (ie in private) is not the issue either. It's the part where the users proceed to publicly share them that poses the problem.
> They didn't make them, nor do they promote them.
By the same logic youtube neither creates nor promotes pirated content that gets uploaded. We have DMCA takedown notices for dealing with precisely this issue.
> Inuitively, Openmetagoog a competitor for manufacturing my IP, and that is also intuitively wrong.
Let's be clear about the distinction between trademark and copyright here. Outputting a verbatim copy is indeed a problem. Outputting a likeness is not, but an end user could certainly proceed to (mis)use that output in a manner that is.
Intent matters here. A product whose primary purpose is IP infringement is entirely different from one whose purpose is general but could potentially be used to infringe.