Commons talk:AI-generated media/Archive 2

From Wikimedia Commons, the free media repository
Jump to navigation Jump to search
← Archive 1 Archive 2 Archive 3 →

AI-improved/restored photographs

Some software can automatically:

  • "Improve" a picture of en:Lake Titicaca using generative AI to transform green areas into realistic trees, transform dots into en:Florida ducks, etc.
  • "Restore" an old black&white picture of en:Léon Blum and colorize it, split hair into individual strands of hair, transform a photo grain fleck into a realistic en:wart somewhere on the skin, etc.

The resulting pictures are works of art/hallucination more than encyclopedic representations of reality. Do we already have a policy page for these? If not, should we draft a policy here?

By the way, I recently raised a concern about the rising trend of camera apps to perform such "improvements" by default without the user even knowing, it might be too early to worry but here it is for possible reference: Commons:Village_pump#Mainstream_cameras_generating_non-existent_artifacts_(AI_generative_upscaling?),_what_can_we_do_about_it?

Thanks! Syced (talk) 06:36, 13 March 2023 (UTC)

"Artificially upscaling or enlarging using any tool, including AI-based or deep learning services" get a mention at COM:OVERWRITE, although only to say that such versions shouldn't overwrite the original. I have sadly seen a few cases of this recently, including bad MyHeritage output that just fakes extra detail onto the eyes, nose and mouth and makes sometimes absurd guesses about facial hair.
Commons:Deletion requests/Files uploaded by Madelgarius is an open discussion about whether to delete a hundred such upscaled portrait photos from Commons. Belbury (talk) 13:26, 13 March 2023 (UTC)
@Syced: Other than the guidance already provided at Commons:AI-generated media#Categorization, I think someone should create a warning template for such cases that can be applied on the File pages. Nosferattus (talk) 18:37, 18 March 2023 (UTC)

Hosting of AI-generated media should follow a whitelist model

that is, uploads are allowed only with prior approval by the community. in some way similar to com:bots, which must be approved before deployment.

the reason is, ai is not only capable of generating images but realistic audio and video. this page and the discussions so far have only considered images. only 1 section above mentions deepfakes, which are one kind of ai videos. ai can do videos much better than deepfakes. ai can also do audio: https://ai.googleblog.com/2022/10/audiolm-language-modeling-approach-to.html https://google-research.github.io/seanet/audiolm/examples/ .

i believe it's not beneficial or ethical to host ai-generated media alongside manual creative works. even if it's allowed, ai-generated media should be clearly tagged as such, and put into a separate category. for example, ai-generated media about cities -- be it images or audiovisual files -- should have its own "category:AI-generated media about cities" under Category:Cities. ai photos should not be mixed with human-made photos. ai audio should not be mixed into Category:Audio files of cities. ai videos should not be mixed with other human-made videos in Category:Videos by city.

we should use a whitelist model so as to deter any unwanted ai-generated media before the situation becomes unmanageable. RZuo (talk) 17:00, 21 March 2023 (UTC)

Hello @RZuo, I agree that AI content should be made identifiable as such, in a suitable way. However, I disagree that all uploads need a prior permission by the community. This would also be difficult, because, how should community members know the image before it is uploaded? "The community" might be only a handfull of people who are really so committed to the cause that they would invest hours in related discussions. / I am not quite sure what do you mean by a whitelist in this context, but maybe I did not understand correctly. Kind regards, Ziko van Dijk (talk) 16:15, 22 March 2023 (UTC)
whitelist.
better deny any upload instead of sieving uploaded files for ai content. it's impossible to do the latter, especially with audio. photos are also super realistic now.
anyone uploading ai content without approval should be indef blocked on sight and their uploads deleted. RZuo (talk) 16:24, 22 March 2023 (UTC)

Uploads by PixelPenguin87

PixelPenguin87 (talk · contribs)

Might wanna contact the legal team on what to do with these type of AI generated images. This sort of photorealistic AI generated content could potentially be a violation of Commons:CHILDPROTECT Trade (talk) 21:46, 18 March 2023 (UTC)

@Brianjd, Ricky81682, and Nosferattus: --Trade (talk) 21:53, 18 March 2023 (UTC)
@King of Hearts: --Trade (talk) 03:28, 19 March 2023 (UTC)
I fully agree. A WMF Legal clarification would be useful. Wutsje 23:06, 18 March 2023 (UTC)
See also U93rFh2T (talk · contribs), VibrantExplorer372 (talk · contribs) and BlueMoon2023 (talk · contribs). Wutsje 03:08, 19 March 2023 (UTC)
Have you tried to seek WMF out for a legal clarification? Trade (talk) 15:49, 21 March 2023 (UTC)
I don't think we need legal advice to decide just to delete this sort of thing. - Jmabel ! talk 16:35, 21 March 2023 (UTC)
I didn't wanted to risk going against community consensus. Trade (talk) 20:16, 31 March 2023 (UTC)

m:Wikilegal/Copyright Analysis of ChatGPT has been published a couple of days ago. It is primarily about text, but it also briefly mentions AI-generated images. whym (talk) 13:54, 24 March 2023 (UTC)

United Kingdom

{{PD-algorithm}} contains a note "The United Kingdom provides a limited term of copyright protection for computer-generated works of 50 years from creation", citing this UK Intellectual Property Office page. From this page:

Unlike most other countries, the UK protects computer-generated works which do not have a human creator (s178 CDPA). The law designates the author of such a work as “the person by whom the arrangements necessary for the creation of the work are undertaken” (s9(3) CDPA). Protection lasts for 50 years from the date the work is made (s12(7) CDPA).

I think it might be tricky to ascertain who "the person by whom the arrangements necessary for the creation of the work are undertaken" might be in modern AI generated media (the author of the prompt? The programmer(s) of the software? Both? - because the software itself is certainly "necessary for the creation of the work", and the prompt as well), but it seems that in the UK, AI generated media is protected anyway, even if we might be unsure who's the owner of the rights. The Office also states there:

The UK remains one of only a handful of countries worldwide that provides this protection. Investment in AI has taken place in other countries, such as the United States, which do not provide this type of protection. Some people argue that this protection is not needed, and others that it should be provided differently.

So, I think that this page should be amended, too, in some way, to contain this information that in the UK and in a "handful of countries" (which countries?) there is protection for AI art, and in deletion discussions, I assume that we have to check where the image was generated - if generated in the US or most other countries, it's {{PD-algorithm}}, but if generated in the UK or the "handful of countries" with protection, we must delete it. In the case of Commons:Deletion requests/File:Alice and Sparkle cover.jpg which I decided to keep, I think it's fine, as the author of the prompt per this article is currently living in California, and was using Midjourney, which is based in San Francisco. Gestumblindi (talk) 20:58, 9 June 2023 (UTC)

It's possible the author didn't knew what other countries he was referring too and simply made a reasonable assumption Trade (talk) 21:05, 9 June 2023 (UTC)
Well, to remedy the lack of UK information, I made an addition based on this UK Intellectual Property Office information. Gestumblindi (talk) 18:24, 10 June 2023 (UTC)

Privacy of living people

Starting a new section to specifically discuss the "Privacy of living people" section. I don't think the current wording reflects the consensus on Commons, but I'm not sure what to change it to. Suggestions? Nosferattus (talk) 17:55, 19 February 2023 (UTC)

At the very least, the claim that "there is a certain likelihood that any AI-generated image with a human face in it violates the privacy of a living person" should be backed up by evidence or removed (also given that other hand-wavy probability claims added by the same user turn out to be highly dubious upon closer inspection, see above). Regards, HaeB (talk) 15:16, 20 February 2023 (UTC)
I am concerned about scenarios like these:
Gnom (talk) 22:19, 22 February 2023 (UTC)
Just to remind everyone: these might be issues in some countries, but basically are not issues in the U.S. (where our servers are hosted) unless the image is somehow disparaging or is used in a way that implies an endorsement by that person. - Jmabel ! talk 22:39, 22 February 2023 (UTC)
@Gnom: The section that discusses privacy issues is under the question "Can AI-generated media be lawfully hosted on Commons?". Can you elaborate on what laws you are concerned with? While I share your general concern for the privacy of living people, we need to focus on what the actual legal issues are (and how Commons typically deals with those legal issues). If your concerns are not legal in nature, they should be moved to a different section. Nosferattus (talk) 00:49, 24 February 2023 (UTC)
On a legal level, this is of course a violation of the GDPR insofar as EU residents are being depicted. Under U.S. law, I am also quite confident that would be illegal to host such images, but others are probably in a better position to assess this. On a non-legal level, I can only say that I would go berserk if an AI were to spit out my face under certain conditions. Gnom (talk) 08:12, 24 February 2023 (UTC)
@Gnom AI-generated art may show faces of living people who were not asked for their consent to appear in the output because photos or drawings depicting them were used to train the software. Accordingly, there is a certain likelihood that any AI-generated image with a human face in it violates the privacy of a living person. Under the Precautionary Principle, these respective files may be nominated for deletion.
No. There are a significant number of countries where consent is not required, and Commons generally does not impose its own consent requirements unless the photo seems to be undignified.
Ironically, the page includes an AI-generated image with a human face in it (Mona Lisa by Midjourney AI.jpg). Should it be deleted under the precautionary principle? Of course not, but there’s nothing on the page to explain this.
Where we can re-use existing laws, guidelines and policies, we should do so. This page should cover only AI-specific issues, and link to other pages for the other issues. Brianjd (talk) 15:02, 25 February 2023 (UTC)
In the US (where Commons is hosted), Australia (where I live), and some other countries, it is generally legal to take and publish images of people in public without their consent. The EU and the US have very different attitudes to privacy, and people living in one of these jurisdictions must be very careful to avoid false assumptions about the other. Brianjd (talk) 15:05, 25 February 2023 (UTC)
On another matter, why focus on faces? Images of people can violate privacy without depicting faces. Brianjd (talk) 15:06, 25 February 2023 (UTC)
Hi, I agree that the Mona Lisa is a bad example (and I should note that was not the one to put it there). I would prefer the examples I inserted above instead. The left image, assuming(!) it were to actually show the face of a real underage girl, would give her parents a right to sue the Wikimedia Foundation for violating their daughter's privacy, even under U.S. law.
Also, you are of course correct that not only the depiction of a person's face can constitute a violation of their privacy, but that is where I see the main issue. Gnom (talk) 20:17, 4 March 2023 (UTC)
There is a problem if an image shows an actual person by the actual original image. But... when the AI creates a new image based on a dozen or thousands of original images, and someone says: "That new image looks like person X!", is that a privacy issue? How decides that the image looks like person X? How to determine that; "looking alike" can be the subjective judgement of an individual. I for example believe that Tim Curry looks like Edmond Privat. Ziko van Dijk (talk) 18:59, 5 March 2023 (UTC)
I would think this is an area where we would not want to host things that are in the gray zones. - Jmabel ! talk 20:49, 5 March 2023 (UTC)
@Ziko The cases that I am thinking about are those where there is simply no question whether or not the AI-generated image depicts a real person. Gnom (talk) 08:08, 6 March 2023 (UTC)
@Gnom, can you give an example of what you mean? In which context would there be no question? Ziko van Dijk (talk) 08:13, 6 March 2023 (UTC)
If the similarity between the AI-generated human face and their real likeness is so close that if you asked 100 people, 99 would say that the two images (the AI-generated image and a photo of the actual human face) depict the same person. That's the scenario I would like to discuss. Gnom (talk) 15:04, 6 March 2023 (UTC)
You vastly overestimate how skillet your average person are at telling the difference Trade (talk) 21:49, 18 March 2023 (UTC)

@Gnom: While I agree with you in principle, I think you have a misunderstanding of U.S. law. In the U.S. there are very very few protections for privacy and strong protections for expression. For example, in the U.S. I could go to a public playground, take photos of a 5-year-old child, ask no one's permission or consent, and then post those photos on the internet or in an art show or whatever. Not only that, but I could even take photos of someone through the window of their home, and in many states it would be perfectly legal. As long as there is no nudity involved (which triggers the federal video voyeurism law) or commercial use involved (which triggers rights of publicity), you can violate people's privacy all day long in the U.S. If you don't believe me, read [1]. This is what happens when you enshrine free speech in your constitution, but not privacy. This lack of privacy protection is also what specifically caused Roe v. Wade to recently be overturned in the U.S. Supreme Court. The Court ruled that because there is no actual "right to privacy" in the U.S., Roe v. Wade was invalid and thus abortion can be prohibited by the law. Of course, I don't believe that Commons policy on privacy should be based on whatever is allowed by U.S. law, but we can't pretend that such a policy is based on the law when it isn't. Our policy should be based on basic ethics and respect for human dignity. But we have to convince the Commons community to agree to such principles in this case. It can't just be taken for granted. Nosferattus (talk) 19:38, 18 March 2023 (UTC)

Since there doesn't seem to be consensus for the contents of this section, I've removed it. I would ideally like to replace it with some sort of guidance we can agree on. Does anyone have suggestions? Nosferattus (talk) 18:46, 19 August 2023 (UTC)

A large-scale AI-generated images upload

Just for the records, in Commons:Deletion requests/Files in Category:AI images created by David S. Soriano I would have decided exactly like Minorax. Per COM:INUSE, it would be hard to argue for deleting files that are actually already in use - and they were added to articles and user pages by various users, not by the uploader -, and this shows that, apparently, these images can be considered useful, but on the other hand, I think there is really a "slippery slope": If we just blanket kept the images, the door would be open for flooding Commons with thousands, tens of thousands - well, why not millions? - of such images, as it's very easy to mass generate this kind of content. Gestumblindi (talk) 09:11, 18 June 2023 (UTC)

Can we know how stable those usages are, beyond the binary in-use vs not in-use? If it has to be checked manually, I hope there is a better tooling to track usage over time. It looks like File:UFO From Distant Past.png was kept for being in use, but it's not in use at the moment. whym (talk) 03:15, 19 June 2023 (UTC)
Given that the remaining files appear mostly on user pages, I'd expect that any mainspace project usage was being washed out days or weeks later. File:UFO From Distant Past.png is the seventh result in a Commons search for "ufo" and certainly the most striking and exciting depiction, but if anyone added it to a project article about UFOs it would (as an illustration of no particular UFO case, drawn in part or in full by a hallucinating AI) be reverted or replaced with something more useful. The same goes for Soriano's cubism uploads: a superficially attentive editor might search Commons for "cubism" and add one to an article as an example of that style, but an editor paying more attention would replace it with a free image by a notable and human artist. Belbury (talk) 10:59, 19 June 2023 (UTC)
This would be far easier to solve if anyone could come in contact with David Trade (talk) 01:38, 27 June 2023 (UTC)

Can the prompts themself be copyrighted?

I am seeing a trend where users have two separate copyright tags. One for the AI prompt they used and one for the AI generated image that came from the output. Should we allow this> Trade (talk) 21:05, 28 June 2023 (UTC)

And plenty of others i cant currently remember. Trade (talk) 01:21, 29 June 2023 (UTC)

Those are probably on the borderline of copyrightability (the first more likely than the second, in my view). Given that in both cases they are specifically disavowing copyright on the prompt and placing it in the public domain, it's hard to see anything objectionable. I think I'd object to someone putting something here, claiming copyright on their prompt, and insisting that using even the prompt requires conforming to a license. - Jmabel ! talk 02:00, 29 June 2023 (UTC)
  • Are there any of these where the prompt is being released under a licence? The examples here are really the opposite of this: the uploader has put the prompt deliberately into [sic] the public domain, to clearly disavow any licensing. Andy Dingley (talk) 02:05, 29 June 2023 (UTC)
I think it's fine if the uploader is explicitly declaring the prompt public domain (or CC0). Other cases would need a closer look. Nosferattus (talk) 01:50, 16 July 2023 (UTC)
If the author have the right to license a prompt as PD then they have the right to license the prompt under as CC-BY-SA. Allowing the former but not the latter is essentially saying that they only own the copyright to the prompt as long as it is convenient for Commons Trade (talk) 02:12, 16 July 2023 (UTC)

Prompts need to be disclosed

It seems to me that the main way AI-generated images are of use is precisely as examples of AI-generated images. To that end, I propose that we should (going forward) require that the file page text for all AI-generated images include, at a minimum:

  1. what AI software was used
  2. what prompt was used

I have no problem with some images being grandfathered in, and of course exceptions need to be made for images that are, for one or another reason, notable in their own right. There might even be a reason for some other exception I'm not thinking of, but I've been seeing floods of low-value AI-generated images, often not marked as such, often involving sexualized images of young women or girls. If those last are coming from prompts that do not ask for anything of the sort, then that itself would be important to document. If, as I suspect, they come from prompts that ask for exactly that, then they really don't belong here, any more than someone's personal photos of themselves and their friends.

And, yes, I am discussing two separate (but related) issues here, disclosure and scope. - Jmabel ! talk 16:53, 28 July 2023 (UTC)

For a starter, we need a category for AI images without prompts so we can keep track of the issue Trade (talk) 18:56, 19 August 2023 (UTC)
(which Trade introduced with Category:AI images generated using unidentified prompts). - Jmabel ! talk 23:55, 19 August 2023 (UTC)

Structured data suggestion

The policy might also want to mention something about how to model the structured data for these images, for instance by using P31 – instance of with Q96407327 – synthetic media (suggested in the Wikidata Telegram group). Ciell (talk) 11:59, 7 January 2023 (UTC)

Question about files that are most likely AI-generated, but without definitive proof

Hello, I was wondering what best practice would be in cases where a file is found online that looks very clearly like an AI-generated image, but the creator either did not specify that they were created with AI or they claimed it to be their own work. Examples include this film poster (notice the pupils and hair), as well as this poster (notice the inconsistent teeth, the bizarre foot, the inconsistent windows on buildings, and the weird birds in the sky). Should we allow them to be uploaded under the presumption that they are AI, or should we assume that they're copyrighted unless they're explicitly stated to be AI? Di (they-them) (talk) 00:16, 20 August 2023 (UTC)

Even if an image contains a lot of obvious AI artefacts, you'd also have to assume that there was no subsequent creative human editing of that image, which seems unknowable. Belbury (talk) 08:02, 20 August 2023 (UTC)
There is exactly zero benefit from trying to do original research on whether or not a image was AI generated without the author admitting so. You are just asking to create a huge copyright mess Trade (talk) 22:38, 20 August 2023 (UTC)

Is the note unnecessary?

The note "Although most current media-generating programs qualify as machine learning and not true artificial intelligence, the term 'artificial intelligence' is commonly used colloquially to describe them, and as such is the term used on this page" seems unnecessary to me. Media-generating programs are indeed artificial intelligence as well as machine learning. Machine learning is considered a subset of artificial intelligence. Artificial intelligence doesn't only refer to LLMs, and includes DALLE-2, Midjourney, etc. Machine learning might be more precise, but AI isn't incorrect. I would like to hear what other people think about this. Chamaemelum (talk) 18:32, 20 August 2023 (UTC)

I agree with your assessment. Machine learning is a subset of AI. Nosferattus (talk) 19:49, 20 August 2023 (UTC)

Real-life

@Trade: you added the parenthetical phrase in "AI fan art of (real-life) fictional characters", which seems oxymoronic to me. How can something be both "real-life" and "fictional"? - Jmabel ! talk 01:52, 21 August 2023 (UTC)

Fictional characters that exist outside of the AI generated art in question Trade (talk) 01:54, 21 August 2023 (UTC)
I hoped the name of the category was enough but unfortunately people keep filling it with images that had nothing to do with fan art. Trade (talk) 01:56, 21 August 2023 (UTC)
To be fair, the description at the top of Category:AI-generated fictional characters doesn't suggest not to. And a lot of categorisation happens based on the category name alone.
Would Category:AI-generated fan art be a useful subcategory to create? Belbury (talk) 09:13, 21 August 2023 (UTC)

AI images of real subjects (aka. "deepfakes")

One subject which this draft doesn't seem to address clearly is the topic of AI images which appear to represent real subjects - e.g. real people, real places, real historical events, etc. These images have the potential to be misleading to viewers, and can cause harm to the project by discouraging the contribution of real images, or by being used as substitutes for real images which are already available.

I'd like to address this as follows. This is intentionally strongly worded, but I feel that it's warranted given the potential for deception:

AI-generated images which contain photorealistic depictions of notable people, places, or historical events have the potential to deceive viewers, and must not be uploaded.

If AI-generated images containing these subjects are used as illustrations, effort should be made to use images which cannot be mistaken for photographs, e.g. by prompting the image generation model to use a cartoon art style.

In a limited number of cases, realistic images containing these subjects may be used as demonstrations of AI image generation or "deepfakes". These images should be watermarked to make it clear to viewers and downstream users of these images that they were machine-generated.

Thoughts? Omphalographer (talk) 23:23, 11 September 2023 (UTC)

A COM:WATERMARK on a demonstration image significantly reduces any constructive reuse of it. Anybody wanting to reuse a notable fake like File:Pope Francis in puffy winter jacket.jpg in their classroom or book should be able to get that direct from Commons.
These images would benefit from prominent warning templates, though, and perhaps an explicit "Fake image of..." in the filenames. Belbury (talk) 08:25, 12 September 2023 (UTC)
Why not just add a parameter to the AI template that can be used to indicate whether or not the image depicts a living person? Trade (talk) 10:51, 12 September 2023 (UTC)
The problem I'm concerned with is reuse of these images outside Wikimedia projects, where the image description certainly won't be available and the filename will likely be lost as well. Photorealistic AI-generated images of recognizable subjects should be fairly rare on Wikimedia projects, and I'm confident that editors can come up with some way of marking them which makes their nature clear without being overly intrusive.
How about the rest? Are we on board with the overall principle? Omphalographer (talk) 19:41, 12 September 2023 (UTC)
Seems reasonable to me. And an alternative to a watermark in the narrow sense would be a mandatory notice in a border under the photo. - Jmabel ! talk 20:10, 12 September 2023 (UTC)
No matter the amount of whistles, alarms and whatnot you put up there will always be someone who cant be bothered to read it before posting the image somewhere else Trade (talk) 15:41, 14 September 2023 (UTC)
Certainly. But if the image itself can tell viewers "hey, I'm not real", then at least it has less potential to mislead. Omphalographer (talk) 17:01, 14 September 2023 (UTC)
In what manner is that not covered by the current AI-image license template + related categories and description? Trade (talk) 22:41, 23 September 2023 (UTC)
Because those do not tend to travel with the image itself when it is reproduced. Indeed, if it the image is used incorrectly within a Wikipedia there would be no indication of that unless someone clicks through. - Jmabel ! talk 22:51, 23 September 2023 (UTC)
Even if you click to expand an embedded image in an article, the full license and disclaimer templates are only visible if you then click through to the full image page. To an unsophisticated user, they might as well not exist. Omphalographer (talk) 00:19, 24 September 2023 (UTC)
So in short the only realistic solution would be a warning template that appears when someone from a Wiki project click to expand the image Trade (talk) 22:04, 13 October 2023 (UTC)
@Omphalographer: I think we should change "notable people" to "actual people" and remove "places" (as that seems overly broad and unnecessary to me). Nosferattus (talk) 23:14, 20 September 2023 (UTC)
We may also want to clarify that this doesn't apply to AI-enhanced photographs. Nosferattus (talk) 23:16, 20 September 2023 (UTC)
Excellent point on "notable people"; I agree that this policy should extend to any actual person, not just ones who cross some threshold of notability.
The inclusion of "places" was intentional. A synthetic photo of a specific place can be just as misleading as one of a person or event; consider a synthesized photo of a culturally significant location like the Notre-Dame de Paris or Mecca, for instance.
AI-enhanced photographs are... complicated. There's no obvious line dividing photos which are merely "AI-enhanced" and ones which begin to incorporate content which wasn't present in the source photo. For instance, the "Space Zoom" feature of some Samsung phones replaced photos of the moon with reference photos of the moon - this level of processing would probably be inappropriate for Commons photos. Omphalographer (talk) 00:11, 21 September 2023 (UTC)
@Omphalographer I think there are some legitimate reasons for creating and uploading photorealistic AI images of places, and less danger that they cause harm. For example, an AI generated image of an ice-free Greenland might be useful for a Wikibook discussing climate change. Sure, it could be misleading if used in the wrong context, but it doesn't worry me as much as AI images of people.
So are you suggesting that all AI-enhanced photographs should also be banned? This will probably be the majority of all photographs in the near future, so I wouldn't support that proposal. Nosferattus (talk) 00:32, 20 October 2023 (UTC)
I'm not suggesting that all AI-enhanced photos should be banned, but that the limits of what's considered acceptable "enhancement" need to be examined. Filters which make mild changes like synthetically blurring the background behind a person's face or adjusting contrast are almost certainly fine; ones which add/remove substantial elements to an image or otherwise dramatically modify the nature of the photo (like style-transferring a photograph into a painting or vice versa) are probably not.
With regard to "places", would you be happier if that were worded as "landmarks"? What I had in mind was synthetic photos of notable buildings, monuments, or similarly specific places - not just any location. Omphalographer (talk) 05:57, 20 October 2023 (UTC)
 Oppose - Very strongly against this proposal, which would be highly problematic for many reasons and unwarranted censorship.
Agree with Belbury on prominent warning templates, though, and perhaps an explicit "Fake image of..." in the filenames - we should have prominent templates for AI images in general and prominent warning templates for deepfake ones...a policy on file title requirements is something to consider. Prototyperspective (talk) 22:11, 13 October 2023 (UTC)
 Oppose per Prototyperspective. I fail to see why this issue is fundamentally different from other kinds of images that could be convincingly misrepresented as actual, unaltered photographs depicting real people, real places, real historical events, an issue that is at least a century old (see e.g. w:Censorship of images in the Soviet Union, Category:Manipulated photographs etc).
That said, I would support strengthening existing policies against image descriptions (and files names) that misrepresent such images as as actual photos, whether they are AI-generated, photoshopped (in the sense of edits that go beyond mere aesthetics and change what a general viewer may infer from the image about the depicted person, place etc.) or otherwise altered. That's assuming that we have such policies already - do we? (not seeing anything at Template:Commons policies and guidelines)
PS regarding landmarks: I seem to recall that authenticity issues have been repeatedly debated, years ago already, in context of Wiki Loves Monuments and related contests, with some contributors arguing that alterations like removing a powerline or such that "ruins" a beautiful shot of a monument should not affect eligibility. I do find that problematic too and would support at least a requirement to clearly document such alterations in the file description.
Regards, HaeB (talk) 01:35, 25 October 2023 (UTC)
 Weak oppose per HaeB. Although I'm sympathetic to the idea of banning deepfake images, I think the proposed wording is too broad in one sense (subjects included) and too narrow in another sense (only addressing AI images). I would be open to a proposal focusing on photo-realistic images of people or events that seem intended to deceive or mislead (regardless of whether they are AI generated or not). Nosferattus (talk) 04:47, 25 October 2023 (UTC)

Wikimedia Foundation position on AI-generated content

The Wikimedia Foundation recently submitted some comments to the US Copyright Office in response to a Request for Comments on Artificial Intelligence and Copyright. Many of the points made by the Foundation will likely be of interest here, particularly the opening statement that:

Overall, the Foundation believes that generative AI tools offer benefits to help humans work more efficiently, but that there are risks of harms from abuse of these tools, particularly to generate large quantities of low-quality material.

File:Wikimedia Foundation’s Responses to the US Copyright Office Request for Comments on AI and Copyright, 2023.pdf

Omphalographer (talk) 20:06, 9 November 2023 (UTC)

I wonder if there was a specific DR that made thw Foundation concerned about low quality spam. Or maybe someone just complained to staff staff? Trade (talk) 23:38, 18 November 2023 (UTC)

Custom template for upscaled images

This page currently advises adding {{Retouched picture}} to upscaled images, which if used without inserting specific text gives a neutral message of This is a retouched picture, which means that it has been digitally altered from its original version. with no mention of the AI nature of the manipulation.

Would it be useful to have a custom AI-upscale template that puts the image into a relevant category and also spells out some of the issues with AI upscaling (potentially introducing details which may not be present at all in the original, copyrighted elements, etc), the way that {{Colorized}} specifically warns the user that the coloring is speculative and may differ significantly from the real colors? Belbury (talk) 08:19, 4 October 2023 (UTC)

Prototyperspective (talk) 09:44, 5 October 2023 (UTC)

I've made a rough first draft of such a template at {{AI upscaled}}, which currently looks like this:

This image has been digitally upscaled using AI software.

This process may have introduced inaccurate, speculative details not present in the original picture. The image may also contain copyrightable elements of training data.

When the template is included on a file page it adds that file to Category:Photos modified by AI per the recommendation at Commons:AI-generated_media#Categorization_and_templates.

Feedback appreciated on what the message should say, and what options the template should take. It should probably always include a thumbnail link to the original image (or an alert that the original is freely licenced but hasn't been uploaded to Commons), and an option to say what software was used, if known, so that the file can be subcategorised appropriately.

It may well be worth expanding this to a generic AI template that also covers restoration and generation, but I'll put this forward for now. --Belbury (talk) 08:21, 12 October 2023 (UTC)

Could you make a template for AI misgeneration? Trade (talk) 23:41, 18 November 2023 (UTC)
Would that be meaningfully distinct from your existing {{Bad AI}} template? Belbury (talk) 19:39, 19 November 2023 (UTC)

Interesting

https://twitter.com/Kyatic/status/1725120435644239889 2804:14D:5C32:4673:DAF2:B1E3:1D20:8CB7 03:31, 17 November 2023 (UTC)

This is highly relevant - thanks, whoever you are! Teaser:

This is the best example I've found yet of how derivative AI 'art' is. The person who generated the image on the left asked Midjourney to generate 'an average woman of Afghanistan'. It produced an almost carbon copy of the 1984 photo of Sharbat Gula, taken by Steve McCurry.

If you don't have a Twitter account, you can read the thread at https://nitter.net/Kyatic/status/1725120435644239889.
Omphalographer (talk) 03:59, 17 November 2023 (UTC)
Hypothetically would we be allowed to upload the AI photo here? Trade (talk) 23:35, 18 November 2023 (UTC)
No? The whole discussion is about how it's a derivative work of the National Geographic photo. Omphalographer (talk) 02:15, 19 November 2023 (UTC)
I couldn't find clear info regarding artworks that look very close to non-CCBY photographs at Commons:Derivative works. This particular image may be fine, it's not prohibited just because the person looks similar to an actual person who was photographed and that photograph was the 'inspiration' to the AI creator.
That AI images look very similar to an existing photograph is overall exceptional and depends on issues with training data, parameters/weighting, and the prompts. Moreover, it's possible that this was caused on purpose to make a point or that they had put an extreme weight on high-valued photographs for cases like this while there are only few images of women from Afghanistan in the training data...more likely though the AI simply does not 'understand' (or misunderstands) what is meant by "average" here.
img2img issues
The bigger issue is that you can use images as input images and let AI modify them according to your prompt (example of how this can be really useful). This means some people may upload such an image without specifying the input image so people can check whether or not that is CCBY. If strength of the input image is configured to be e.g. 99% the resulting image would look very similar. I think there should be a policy that when you upload a AI-generated image via img2img, you should specify the input image. Prototyperspective (talk) 11:22, 19 November 2023 (UTC)
If a human had created that image, we would certainly delete it as a plagiaristic copyvio. I see no reason to treat it more favorably because the plagiarist is a a computer program. - Jmabel ! talk 18:48, 19 November 2023 (UTC)
I don't think so. I don't see much use of discussing this particular case and only meant to say that the derivative works does not really have info on this but I think artistic works that show something that has previously been photographed are allowed. Or are artworks of the Eiffel tower not allowed if the first depiction of it is a photograph that is not CCBY? Prototyperspective (talk) 18:58, 19 November 2023 (UTC)
COM:BASEDONPHOTO is the relevant Commons policy for a drawing based on a single photograph: it requires the photographer's permission. I too would see no copyright difference between a human sketching a copy of a single specific photograph and an AI doing the same thing digitally. Belbury (talk) 19:19, 19 November 2023 (UTC)
Thanks, a link to this policy was missing here so far. I don't see how the issue of photographing things like the Eiffel tower to copyright them away is addressed though. In this case, a person was photographed. I object to the notion that if the first photograph of a person, animal, object, or whatever is not in the public domain, it also can't be drawn under CCBY. I too would not see a copyright difference between a human sketching of a single photograph and an AI doing the same thing digitally. That is what img2img is, the case above is not based on a single image but many images, including many images of women. It would never work if it was just on one image. Prototyperspective (talk) 22:01, 19 November 2023 (UTC)
The case above is not based on a single image but many images, including many images of women... I'm not convinced. The results shown look very much like they are based primarily on the National Geographic photo, possibly because there were many copies of it in the training data. Omphalographer (talk) 22:18, 19 November 2023 (UTC)
@Prototyperspective: "the notion that if the first photograph of a person, animal, object, or whatever is not in the public domain, it also can't be drawn under CCBY." That's a straw-man argument, as is your Eiffel Tower example. You are refuting a claim that no one is making. While we have little insight into the "mind" of a generative AI system, I think we can reasonably conclude that if the AI had not seen that particular copyrighted image or works derived from it, then the chance is vanishingly small that it would have produced this particular image. And that is the essence of plagiarism. - Jmabel ! talk 01:06, 20 November 2023 (UTC)
You may be misunderstanding COM:BASEDONPHOTO. It isn't saying that once somebody takes a photo of a subject, they have control over anyone who chooses to sketch the same subject independently in the future. It only applies to someone making a sketch using that photo alone as their reference for the subject.
The "many images" point doesn't seem very different from how a human would approach the same copying process. The human would also be applying various internal models of what women and hair and fabric generally look like, when deciding which details to include and omit, and which textures and styles to use. It would still result in a portrait that had been based closely on a specific existing one, so would be infringing on the original photographer's work. Belbury (talk) 13:01, 20 November 2023 (UTC)
Now there are many good points here.
I won't address them in-depth or make any statements in regards to whether I agree with these points and their conclusion. Just a brief note on an issue & unclarity: what if that photo is the only photo of the organism or object? Let's say you want to draw an accurate artwork of an extinct animal photographed once where you'd orient by/use the photo – I don't think current copyright law finds you are not allowed to do so. In this case, I think this photo is the only known photo of this woman whose noncopyrighted genetics further emphasize her eyes making a certain noncopyrighted facial expression. Prototyperspective (talk) 13:18, 20 November 2023 (UTC)
@Prototyperspective: I am going to assume good faith, and that you are not just arguing for the sake of arguing, but this is the last time I will respond here.
  • If there is exactly one photo (or other image) of a given organism or object, and it is copyrighted, and you create a work that is clearly derivative of it in a degree that is clearly plagiaristic, then most likely you are violating copyright. Consider the Mona Lisa. We don't have any other image of that woman. If it were a painting recent enough to still be in copyright, and you created an artwork that was nearly identical to the Mona Lisa, you'd be violating Leonardo's [hypothetical] copyright.
  • For your "extinct animal" case: probably the way to create another image that did not violate copyright would be to imagine it in a different pose (based at least loosely on images of a related species) and to draw or otherwise create an image of that. But if your drawing was very close to the only known image, and that image was copyrighted, you could well be violating copyright.
  • Again: the user didn't ask the AI to draw this particular woman. They asked for "an average woman of Afghanistan," and received a blatant plagiarism of a particular, iconic photo. Also, you say, "I think this photo is the only known photo of this woman." I suppose that may be an accurate statement of what you think, but it also tells me you have chosen to listen to your own thoughts rather than do any actual research. It is not the only photo of Sharbat Gula, nor even the only published photo of her. Other photos from that photo session when she was 12 years old were published (though they are less iconic) and I have seenn at least two published photos of her as an adult (one from the 2000s and one more recent). I suspect there are others that I have not seen. {[w|Sharbat Gula|She's had quite a life}} and now lives in Italy.
Jmabel ! talk 21:16, 20 November 2023 (UTC)
This exact issue is described in Commons:AI-generated media#Copyrights of authors whose works were used to train the AI. It isn't discussed in other Commons policies because those documents were generally drawn up before AI image generation was a thing. Omphalographer (talk) 19:42, 19 November 2023 (UTC)
Similarly to the issues presented in the Twitter/X thread, there is a lawsuit of a group of artists against several companies (incl. Midjourney and Stability AI), where a similar concern is presented (i.e. AI-generated media taking precedence above directly related images and works). I think this, among other things, is important to consider when deciding what scope Commons has in regards to AI-generated media. EdoAug (talk) 12:57, 9 December 2023 (UTC)