44

Feb 22 2024 Update: I have just rolled out the network-wide version of the help center article that mentions content generated by generative artificial intelligence tools must be referenced. I have overridden this article locally to replace it with the variant that mentions that content is not allowed for the following sites:

  • Ask Ubuntu
  • Unix & Linux
  • Academia
  • Stack Overflow em Português
  • Mathematics
  • Home Improvement
  • Politics
  • SciFi & Fantasy
  • Stack Overflow en Español
  • Software Engineering
  • Super User
  • Worldbuilding
  • Motor Vehicle Maintenance & Repair
  • Latin
  • Writing
  • Programming Language Design and Implementation
  • DBA
  • Personal Finance & Money
  • Retrocomputing
  • Board & Card Games
  • Code Review
  • Movies & TV
  • Arts & Crafts
  • Pets
  • Chemistry
  • Stack Apps
  • MSE

Work left to do:

  • Coordinate with international Stack Overflow sites so ensure both variants have translations, for future use (and assist with guiding the discussion for any of those where the topic hasn't come to the community's attention yet);
  • Make further changes to the referencing Help Center article, as per this answer
  • Changing the copy on the "must be referenced" banner from "cited" to "referenced;"
  • Change the slug for the help center article from ai-policy to gen-ai-policy; change the link in the banner to point there; put appropriate redirects in place.

Feb 16 2024 Update: Please see the revised proposed copy for the relevant help center articles below, that takes your feedback into account ^_^


Feb 14 2024 Update: Thanks for all the feedback. To make sure we have a chance to go through as much of that feedback as possible, we're moving the roll out date from today (Feb 14) to early next week (Feb 20).


In this recent post about the two possible banner options that Stack Exchange sites can adopt to educate answerers about their policy on answers generated by artificial intelligence tools, I mentioned I’d make a separate post to work on a draft for a Help Center article. As such, I’m here to present proposed changes to an existing Help Center article, proposed copy for a new one — and to get your feedback on both of those.

Why are we doing this, again?

As a reminder, and a summary of the proposed timeline:

  • As per our Code of Conduct, any content posted that is generated with artificial intelligence tools must be referenced;
  • On February 20th, a new help center article will go live network-wide at /help/ai-policy, which will reference the aforementioned network-wide policy that content generated by artificial intelligence tools needs to be referenced. Sites that have — before February 20th — made the Community Management Team aware (via the usual system) of their superseding site policy on the matter that disallows this content instead of requiring it to be referenced will have a different template that instead reflects that policy.
  • On the same date, this article will be changed network-wide, to explicitly mention that content generated by artificial intelligence tools are covered by our referencing guidelines, if there isn't a superseding site policy that disallows it altogether. It will have a link to the article mentioned in the bullet point above (the one that lives at /help/ai-policy);
  • No site will have either banner on by default. After the roll out, sites still have the option of changing their policy on content generated by artificial intelligence tools, and the moderation team will be able to edit the Help Center article at /help/ai-policy. They will also be able to request that the Community Management Team enable the banner that reflects their stance on content generated by artificial intelligence tools (to do so, the community on that site must reach a consensus on that decision, as well as on the language that should be present in the Help Center Article that explains that site’s policy on this matter, if it differs from the templates we collaboratively agree on in this post).

The drafts

Below are the proposed changes to the existing Help Center article on referencing:

Plagiarism - posting the work of others with no indication that it is not your own - is frowned on by our community, and may result in your content being downvoted or deleted. This includes content generated by humans, as well as generative artificial intelligence tools if $SiteShortName allows the latter to be posted — please check the $SiteShortName policy on content generated by generative artificial intelligence tools.

Below are the proposed copies (both drawing from the existing article on Stack Overflow as a starting point) for the Help Center article that will live at /help/ai-policy, with the proposed title of “What is this site’s policy on content generated by generative artificial intelligence tools?". Sites will, by default, have the top version, that refers to the network-wide standard that content generated by generative artificial intelligence tools must be referenced:

Generative artificial intelligence (a.k.a. GPT, LLM, generative AI, genAI) tools can be used to generate content for $SiteName, but this content must be properly referenced as per our guidance. If your content is determined to have been written by generative artificial intelligence tools and is not properly referenced, it will likely be deleted, along with any reputation earned from it. Posting unreferenced AI-generated content may lead to a warning from moderators, or possibly a suspension for repeated infractions.

Some sites on the Stack Exchange network may have different policies on content generated by generative artificial intelligence tools, and some may disallow the posting of AI-generated content entirely. Please ensure you check each site’s local policy about this content before posting there.

What counts as “content generated by generative artificial intelligence tools”?

“Content generated by generative artificial intelligence tools” is any content crafted, in part or in whole, using a tool that writes a response automatically based on a prompt it is provided. These tools include large language models like ChatGPT and Google Gemini. Because these tools are trained to answer with language that mimics authentic speech, the responses may look and sound plausible, but the quality of generated answers can vary significantly (up to, and including, completely wrong answers).

If you are using large language model (LLM) services as described above to draft content for $SiteName, please ensure these are properly referenced.

Why do I need to disclose that I used generative artificial intelligence services to draft content?

Stack Exchange is a collaborative resource, developed and maintained by members of the community, with the goal of creating a repository of high-quality curated knowledge. As mentioned above, while content generated by generative artificial intelligence tools may look and sound plausible, the quality of that content can vary significantly. For this reason, it is extremely important that you ensure future readers are aware of the fact that these tools or services were used to generate the content you’re posting. This ensures that readers of that content are fully aware of the content source’s limitations and shortcomings, and can thus decide how relevant that content is for them.

There are a few primary issues with content generated by large language models that make it unsuitable for use on $SiteName without properly referencing it:

  1. Users who ask questions on $SiteName expect to receive an answer authored and vetted by a human. This ensures that the answer is factual, relevant, and complete, up to the standards of another human. While human authors are not perfect, generative artificial intelligence tools may not take into account other important factors that add nuance to a question, often add excessive noise to their answers (e.g., explaining all details, no matter how relevant), and may fabricate false or misleading information.
  2. Users who ask questions on $SiteName may have already sought answers elsewhere, including through generative artificial intelligence services. Appropriate references help the reader decide whether or not they even want to read a given post, and flags the potential hazards involved in posts that have been written while using generative artificial intelligence tools.
  3. Sometimes generative artificial intelligence tools may be used to assist with editing and translating content, rather than generate it. Despite this constituting a more transformative, rather than generative action, these tools are still prone to error, and they might still introduce new information not present in the original text.
  4. Generative artificial intelligence tools are not capable of referencing the sources of knowledge used up to the standards of the Stack Exchange network. Even when generative artificial intelligence tools appear to reference sources for responses, such sources may not be relevant to the original request, or may not exist at all. For $SiteName, this means the content may not honestly or fairly represent the sources of knowledge used, and that the actual original author of some of the material being used may not be getting properly credited, even if someone explicitly references the generative artificial intelligence tool as an author in their answer.

Please note that even when properly referenced, frequent posting of content generated by these tools, especially when done in a short window of time, can lead to unnecessary disruption of $SiteName, which would be in violation of the Inauthentic Usage policy of our Code of Conduct.

How do I reference content generated by generative artificial intelligence tools?

The more general guidance offered in the context of referencing material written by others applies to content generated by generative artificial intelligence tools. More specifically, there are two main things you should ensure:

  • You should clearly describe what content is generated by generative artificial intelligence tools, and which isn’t. You should do this when quoting directly from the output produced by these tools, as well as when paraphrasing those contents. This ensures there is a distinction between the content generated by these tools, content you authored, and content you may be referencing from other sources.
  • You should specify the specific generative artificial intelligence tool/service you used. Since different services may produce different outputs to the same prompt, and may have different limitations and shortcomings, you should ensure readers know which specific tool you used to produce the content you are referencing.

You may want to share details on the prompt you used to produce the output too, but there is no need to completely copy and paste the prompt and output. Just as when referencing any other content you didn’t author, you should generally avoid copying the complete text, and should instead use the words and ideas from this content to support your own. Here is how you might consider referencing material generated by generative artificial intelligence tools in the body of your post:

I asked [Generative AI service] about [partial prompt]. Its output was:

[Generative AI response]

[other sources, quotes, explanations, etc. necessary to complete the post]

If you have a question about how this policy might look in practice, please look through Meta $SiteName to see if there are any past or ongoing discussions about this type of content that establish community norms surrounding it. If there aren’t any such discussions that address your concerns, consider starting a discussion yourself so other members of the community can provide their perspective.

Are there alternatives to using generative artificial intelligence services to craft answers?

Many of the answers on $SiteName are created by users sharing their own expertise. In particular, when supplying answers outside their expertise, users should reference trustworthy sources. Searching for sources, synthesizing them into a good answer, and referencing them clearly are critical steps in developing a well-crafted answer.

Remember - the person who asked a question needs a correct answer. As such, answering correctly is always more important than answering quickly. Additionally, our system thrives on questions getting multiple good answers, which are more likely to help future visitors who have the same question. By following a process that creates consistently correct answers of good quality, and that are well-referenced as needed, you should do just fine here.

And sites that reach community consensus on having a policy that disallows content generated by generative artificial intelligence tools can use the following template as a starting point, so there is some consistency in the language across network sites:

Generative artificial intelligence (a.k.a. GPT, LLM, generative AI, genAI) tools may not be used to generate content for $SiteName. The content you provide must either be your own original work, or your summary of the properly referenced work of others. If your content is determined to have been written by generative artificial intelligence tools, it will likely be deleted, along with any reputation earned from it. Posting content generated by generative artificial intelligence tools may lead to a warning from moderators, or possibly a suspension for repeated infractions.

What counts as “content generated by generative artificial intelligence tools”?

“Content generated by generative artificial intelligence tools” is any content crafted, in part or in whole, using a tool that writes a response automatically based on a prompt it is provided. These tools include large language models like ChatGPT and Google Gemini. Because these tools are trained to answer with language that matches authentic text, the responses may look and sound plausible, but the quality of generated answers can vary significantly (up to, and including, completely wrong answers).

Please do not draft content for $SiteName using large language model (LLM) services as described above.

Why am I not allowed to use generative artificial intelligence services to draft my content?

Stack Exchange is a collaborative resource, developed and maintained by members of the community. There are a few primary issues with content generated by large language models that makes it unsuitable for use on $SiteName:

  1. Users who ask questions on $SiteName expect to receive an answer authored and vetted by a human. This ensures that the answer is factual, relevant, and complete, up to the standards of another human. While human authors are not perfect, generative artificial intelligence tools may not take into account other important factors to a question (e.g., optimization, security, etc.), often add excessive noise to their answers (e.g., explaining all details, no matter how relevant), and may fabricate false or misleading information.
  2. Users who ask questions on $SiteName may have already sought answers elsewhere. Due to the ease of using generative artificial intelligence services, if a user wanted an answer from an artificial intelligence, they may already have sought one, and so it does not make sense to provide one here.
  3. Generative artificial intelligence tools are not capable of citing the sources of knowledge used up to the standards of the Stack Exchange network. Even when generative artificial intelligence tools appear to cite sources for responses, such sources may not be relevant to the original request, or may not exist at all. For $SiteName, this means the content may not honestly or fairly represent the sources of knowledge used, even if someone explicitly cites the generative artificial intelligence tool as an author in their content.

Are there alternatives to using generative artificial intelligence services?

Many of the answers on $SiteName are created by users sharing their own expertise. In particular, when supplying answers outside their expertise, users should reference trustworthy sources. Searching for sources, synthesizing them into a good answer, and referencing them clearly are critical steps in developing a well-crafted answer.

Remember - the person who asked a question needs a correct answer. As such, answering correctly is always more important than answering quickly. Additionally, our system thrives on questions getting multiple good answers, which are more likely to help future visitors who have the same question. By following a process that creates consistently correct answers of good quality, and that are well-referenced as needed, you should do just fine here.

Feedback

We’ll keep this post open for community feedback until February 19, 2024, and will be responding to feedback during that period. Once we collectively land on the copy to be used on these two Help Center articles, I’ll update the original post about the banners with next steps on rolling this out network-wide. Thanks for your help! ^_^

8
  • 1
    If moderators on each site can override this to have whatever policy they want (if I understood the situation correctly) then whats the point of even having this? I would assume if this is the entire site's policy, all sub-sites would have a policy that would be a subset of this and not some completely different random thing. And whatever happened to the heuristics that a post is supposed to have met, before being actionable for it being created using help from an ai tool?
    – user13267
    Commented Feb 8 at 1:56
  • 7
    The default, network-wide stance on AI-generated content is that, if allowed on any site, it needs to be properly cited; some sites may choose to completely disallow it instead. I'm laying out a framework that allows sites to decide which of the two they want, and to request a corresponding banner be enabled. The heuristics will apply regardless of the site policy, as I see it — they serve for folks to determine whether the content "is AI-generated" or not; the action taken once the determination is made varies depending on the policy. Can you clarify what your concern is, @user13267?
    – JNat StaffMod
    Commented Feb 8 at 15:08
  • 1
    What if a site already has a "No AI Answers" policy?
    – T.E.D.
    Commented Feb 8 at 18:47
  • 2
    Just signal it to the CM Team that you don't want the banner enabled, @T.E.D. ;) (it's the bullet point in bold near the top of the post)
    – JNat StaffMod
    Commented Feb 8 at 19:08
  • 4
    Apologies for changing the roll-out plan, and for any confusion this might've caused. Revision 9 of the post lands on (hopefully!) the final form of this plan: all sites will have a new help center article explaining why AI-generated content needs to be cited, unless they have a superseding policy that disallows it. In discussion with mods in the TL, it became apparent that the "needs citation" banner shouldn't be on by default, and instead any banner should only be enabled by request from that community.
    – JNat StaffMod
    Commented Feb 8 at 19:51
  • 2
    As per suggestions in one of the answers, I'll also be providing a second template to be used for sites that opt for a policy that disallows this content, so they don't need to start their help center articles from scratch. Thanks for your patience, feedback, and help, everyone! ^_^'
    – JNat StaffMod
    Commented Feb 8 at 19:52
  • If there isn't going to be a clear universal policy with regards to ai tool usage for all se sites, then does SE have any plans to clarify if the heuristics are being followed properly ? General info on the status of how (or if) they are being used?
    – user13267
    Commented Feb 9 at 5:43
  • 1
    The edits seem to address my concerns with the original versions. I would propose one minor change: "may be used to assist with translating content" -> "may be used to assist with editing content" or "may be used to assist with editing and translating content". Translating alone is likely to be too narrow and not inclusive of other types of editing. Commented Feb 16 at 15:08

9 Answers 9

30

Revision 8 of the question addresses these concerns.


Answers are not the only part of the content that can be used as knowledgebase and where posting AI content can do harm we are poorly equipped to deal with, besides relying on AI policy. On Stack Overflow there are also articles and discussions.

Instead of using word answers, use word posts in AI policy text. Or at least explicitly name articles and discussions there, too.

When warning users about posting content other than answers linking to AI policy that explicitly mentions answers leaves plenty of room for interpretation. It may look like the answers are the only parts where AI is not allowed.

Similar request has already been made at Stack Overflow Meta Extend AI content policy to articles and other content that is meant to be used as part of the knowledgebase

7
  • 7
    Not again!
    – Laurel
    Commented Feb 7 at 21:42
  • 2
    Or even posts -> the more generic content, because AI-generated content is the same level of dangerous anywhere it can be pasted - e.g. reviews and comments. Even comments can persist as part of the knowledge base. When allowed on a site, it should also be properly cited - even in comments. Commented Feb 8 at 15:11
  • Even the bio, as some people have put answers to common questions in their about me section, and sometimes direct people to their bio for that answer. I don’t have an explicit example handy, this is just memory. I haven’t seen anyone post AI content there, but if they can… Commented Feb 8 at 15:16
  • 2
    At least on History.SE, using AI in questions is a whole different kettle of fish. I don't think we have an issue with that in theory, but in practice what we see is people asking questions of the ilk of ("Which is right, this professional historian, or this random crap spit out by an AI?"), and we don't want to allow those.
    – T.E.D.
    Commented Feb 8 at 18:55
  • @T.E.D. We have those on Stack Overflow, too. Some are "please debug the code AI spit out and I don't understand", but some are genuine questions that use AI for rephrasing. We can deal with first ones as regular poor questions, and I personally don't see too much harm in second kind, besides the fact that people who tend to use AI for questions will use it in other places. Also AI tends to write a lot more than strictly necessary, so I cannot really say that such questions are really improved comparing to how would they look like without AI help. Overall, I don't see AI is useful in any way. Commented Feb 8 at 19:04
  • 1
    @ResistanceIsFutile - Well, our issue is I think more subtle, and a better comparison would be the "notability" requirements for questions on Skeptics. We just feel that information from an AI is inherently unreliable, so it cannot be used to establish the kind of confusion of the facts that would necessitate a question in the first place.
    – T.E.D.
    Commented Feb 8 at 19:07
  • @T.E.D. Yes, I definitely see how can that be a problem. Commented Feb 8 at 20:03
21

Here's my thoughts:

  1. The proposed wording is this:

    Posting unreferenced AI-generated content repeatedly may lead to a warning from moderators, or even a suspension from the site.

    While it's reasonable to wait to issue a suspension until repeated violations have occurred and the user has been warned, I think that (ideally) users should be warned on the initial rule violation. I don't see much value in waiting until there's multiple rule violations to tell a user to stop.

    Perhaps something like this:

    Posting unreferenced AI-generated content may lead to a warning from moderators, or possibly a suspension for repeated infractions.

  2. Can there be a non-English version of this policy for non-English SE sites?

  3. Can (should?) there be standardized phrasing for a ban on AI content, whether cited or not? If not, individual sites have to create that themselves. While that can work, a template for that too would be nice.

  4. It states this:

    Some examples of AI services commonly used to write infringing answers on $SiteName include [...]

    While this may be true on large sites, like Stack Overflow, can this say "Some examples of AI services commonly used to write infringing answers include [...]" instead? It's quite possibly smaller sites haven't seen AI answers from a variety of services.

  5. The article alternates between "AI" and "artificial intelligence" a number of times. I'd probably suggest not switching constantly. Stating "AI (artificial intelligence)" and then just using "AI" might be better.

  6. Google Bard is now Gemini. It may be a good idea to update the phrasing to use "Google Gemini" or perhaps "Google Gemini (formerly Google Bard)". (I'm not endorsing a specific news site here, other news sources have

7
  • 1
    for 3., I'm not sure what you mean: you're asking for a template for what context, specifically?
    – JNat StaffMod
    Commented Feb 8 at 14:56
  • @JNat Well - advice on citation requirements for genAI posts are pointless if its completely banned. "tools can be used to generate answers for $SiteName" would be irrelevant if we'd rather not have people do so Commented Feb 8 at 15:13
  • 1
    ah, got it: a template for sites to use as a starting point for their help center article if they choose to disallow the content, rather than require it be cited! That makes complete sense, yes.
    – JNat StaffMod
    Commented Feb 8 at 15:15
  • 1
    revision 7 addresses your point 5.
    – JNat StaffMod
    Commented Feb 8 at 17:54
  • 1
    (not ignoring the other points; still plan to go back to them)
    – JNat StaffMod
    Commented Feb 8 at 19:54
  • @JNat No problem, thanks for looking at it! FYI: I've added one more suggestion as it seems like Google Bard is being renamed.
    – cocomac
    Commented Feb 8 at 21:38
  • 2
    For 2., I'll be working with the International SO sites on coming up with translations for the two variants of the articles ^_^
    – JNat StaffMod
    Commented Feb 16 at 15:04
14

TL;DR Please replace the word "cite" with "reference" or "attribution" throughout.

As per our Code of Conduct, any content posted that is generated with artificial intelligence tools must be cited;

This is not what the Code of Conduct says.

Citing is not the same as referencing or acknowledging

— see revision 13 for changes that address this concern

In a science community, citing means to reference a publication. The publication does not change over time, and unless there is a global catastrophe, you will be able to find the publication, even if it was written a hundred years ago, in multiple libraries or databases. In modern times, many citable publications have a document object identifier, DOI. For example, DOI:10.1038/171737a0 is Watson's and Crick's 1953 paper on the double-helical model of DNA. Not the draft, not the version updated in 1967, not a version copy-edited by AI, but the published final version as printed.

The Code of Conduct, in the inauthentic usage section, says that plagiarism is not allowed, and links to referencing standards:

If you copy (or closely rephrase/reword) content that you did not create into something you post on Meta Stack Exchange (e.g., from another site or elsewhere on Meta Stack Exchange), make sure you do all of the following:

  • Provide a link to the original page or answer

  • Quote only the relevant portion

  • Provide the name of the original author

This ensures that the original creator gets credit for their work.

This is not how you cite published sources, this is how you reference or attribute all kinds of information, a less rigorous standard. It is a bit like saying "(Prashanth Chandrasekar, personal communication)".

"This ensures that the original creator gets credit for their work."

— see revision 13 for changes that address this concern, at least partially

If you reference text generated by AI, even in the most rigorous way (by providing a (perma-)link to a document that contains both prompt and generated material), the original creators of the work do not get credit. This is different from a scientific paper you cite because when it relies on the work of others, it will in turn cite sources that were used, giving a network of authors involved in building this knowledge. When you post AI-generated text, this chain of evidence gets disrupted, and you have assume responsibility for the posted text (and check whether it makes sense, and find authentic sources to support the claims it makes).

The example in the referencing help text

The example quotes a text from a web site. The web site does not give the author of the text. It references two works, but the links are dead and get rerouted. It would be nice to update the example so that it follows the referencing standard.

The definition of plagiarism

Plagiarism is not defined within the text, but instead links to the article on wikipedia. This article will change with time, so it would have been better to link to the version of the article at the time the help text was written. As is, we define plagiarism outside of StackExchange, but use the term in the code of conduct. Importantly, the outside definition may change to include text generated by AI, or stick with the more common text written by someone else (another human, someone who deserves credit).

It says "plagiarism is frowned on" but the Code of Conduct says it is not allowed on the site.

Why is this so hard

We are trying to make rules for a technology that is in its infancy. Today, ChatGPT might not be able to connect answers with sources, but in a couple of months the next iteration might. Right now, ChatGPT is horrible at math, but the GPT4/Wolfram combo is already pretty good. Also, different people have different standards. It is the job of the community to ensure that there is a single well-explained standard if that standard is used to expel community members who do not follow it.

5
  • Rather than use an example of academic dishonesty to point out an academic dishonesty-related jargon issue, you could use something other than Watson and Crick.
    – Nij
    Commented Feb 9 at 4:36
  • @Nij There is some irony in that, I agree.
    – Karsten
    Commented Feb 9 at 5:28
  • 1
    I'm confused about what the issue with the example in the referencing article is. The example quotes from a page; the help center links to the page where it quotes from. The fact that that page doesn't mention who wrote that text, and that it has links we would normally not accept as direct references as per our standards seems secondary — the fact remains that the relevant piece of quoted text is properly referenced as per our standards. If we wanna change the example to one without the redirects/dead links, and with a clearer author, we can; but the current example is as per standard, no?
    – JNat StaffMod
    Commented Feb 14 at 15:02
  • @JNat The current example does not seem to follow two items in the standard, "Provide the name of the original author", and "This ensures that the original creator gets credit for their work." I don't know who wrote the page, and so the original creator does not get credit for their work. Giving a link to a page is always a problem because 10 years later, the URL might be gone or the content might have changed. If you link to a Wikipedia article (like for the definition of plagiarism), the page is very likely to be different after 10 years, unless you link to a specific version of the article.
    – Karsten
    Commented Feb 15 at 2:00
  • 2
    For the parts of your answer marked as [status-planned] as of the answer's third revision: we plan to work on these, but feel these changes aren't necessarily blockers for the other proposed changes, so we'll be working on them after rolling out the other changes currently proposed in revision 13 of the question.
    – JNat StaffMod
    Commented Feb 16 at 14:57
9

See revision 5 of the question for proposed rephrasing to address the concerns raised in this answer.


Please check what the policy on content generated by artificial intelligence tools on $SiteShortName is, since this guidance on referencing may apply to those, in addition to content generated by humans.

The wording of "...may apply" indicates that it's possible that a site could allow uncited AI-generated content, which is misleading; it requires attribution at a bare minimum on every site on the network, via the Code of Conduct.

I'd personally opt for language along the lines of

Plagiarism - posting the work of others with no indication that it is not your own - is frowned on by our community, and will result in your content being edited or deleted. This includes content generated by artificial intelligence tools, which require disclosure on the Stack Exchange Network. Please check the specific policy on $SiteShortName regarding AI-generated content for further information.

This makes it clearer that it's not just "encouraged" to provide attribution, but required, and violations will result in action being taken on the content. It then directs you to the specific policy on the site in question.

3
  • 2
    The "may apply" is there because it may be the case that a site completely disallows AI generated answers — and in that case, it really doesn't apply. Hence the link to the actual policy on AI content afterwards. I'm not aiming to have an article that encourages AI content to be cited, but rather to point out that the site's actual policy on AI content lies elsewhere, and depending on what it is it's possible these guidelines on referencing apply to it to. Does that make sense?
    – JNat StaffMod
    Commented Feb 8 at 14:49
  • 2
    Please see revision 5 for proposed rewording, which I think addresses the concerns.
    – JNat StaffMod
    Commented Feb 8 at 16:18
  • 1
    gonna consider this [status-completed] unless you come back to this and tell me my changes don't address your concerns ^_^
    – JNat StaffMod
    Commented Feb 13 at 15:36
9

— see revision 13 for changes that address this concern


Because this is a late addition, I'm adding it as a second answer for visibility purposes.

In addition to plagiarism, the Inauthentic Usage policy talks about "directly causes unnecessary and unwanted disruption ... to users and/or the network". I believe that it's worth calling this out in the AI policy Help Center documentation somewhere. A pattern that we've seen on multiple sites is users flooding the site with AI generated content. Being able to quickly generate and post content that looks meaningful is even more disruptive than someone performing destructive edits or flooding with spam or nonsense because it can be a little less obvious. Even on sites that do allow AI generated content with attribution, people (especially those with enough reputation to bypass some rate limits) should take care to not post too frequently.

4

See revision 11 of the question.


please check what the specific policy on content generated by artificial intelligence tools on $SiteShortName is.

This is wordy and confusing. I recommend instead:

please check the $SiteShortName policy on content generated by artificial intelligence tools.

or:

please check the $SiteShortName-specific policy on content generated by artificial intelligence tools.

4

Factually, equating Bard to LaMDA is wrong. LaMDA was an early Google large language model. Today, Bard uses Gemini. Between LaMDA and Gemini, Google also had PaLM. Also, as of 8 February 2024, Bard no longer exists - several Google offerings, such as Assistant, Bard, and Duet AI in Google Workspace, have been rebranded as Gemini.


— see revision 13 for changes that address this concern

I would also recommend clarity on the use of AI tools as editors. As we see these technologies translate into tools like translation services or editors (tools like Grammarly or Microsoft Editor), we should have considerations that these tools are more likely to be allowed. With my understanding of the heuristics that are available to moderators to guide judgement calls on if a post is AI or not, I think these tools would generally be OK. They tend to be more transformative than generative in nature but aren't always free from error. However, assuming they aren't generating new content, they can be more useful in improving the accessibility of the site and its content to a broader range of people, especially on the English-only sites in the network.


— see revision 13 for changes that address this concern. Guidance that needs to be more specific than the one proposed is probably better decided on a per-site basis, and changes are made locally to that site's help center article.

The edge case for questions about the use of generative AI also needs to be considered. Even on sites that prohibit the use of AI to generate answers, an answer including generative AI output is likely not prohibited if that output supports the answer. On Software Engineering, for example, I can see questions about the use of generative AI technologies to support requirements engineering. An answer that includes information about prompting techniques and examples of output (properly attributed) would be allowed, assuming that the rest of the answer was written by a human and the generated portions supported the rest of the answer.


— see revision 13 for changes that address this concern

I'm also struggling to reconcile the points made in the answer to why disclosure is needed with the continued stance that prohibitions on AI-generated content should be site-by-site rather than network wide. The three risks presented are presented in a way that is very compelling for outright prohibition of generated content. I would expect that many, if not most, reasonable people would look at those risks and conclude that curating generated content, even with attribution, is not worth the risks of noisy, false, misleading, unwanted, or otherwise low quality answers.

9
  • 1
    Upvoting for your second point; I think the wiki really needs to at least briefly touch on the use of AI editing tools. I don't think it should necessarily go as far as specifying they're allowed or disallowed as much as it should be something like "use caution, the output is your responsibility to verify".
    – zcoop98
    Commented Feb 8 at 0:24
  • 3
    One of the issues with AI editing and translation tools, is that they are not merely fixing grammar or translating, they are rewriting content in a way that follows general AI heuristics. This makes it harder to detect between post that is written by a person and then run through such tool or completely generated. It is also used as frequent excuse by people who were definitely using AI for generating whole post (because of other visible heuristics) claiming they were just using it for translation. As long as we have people running rampant with posting AI, I would prefer those are banned, too Commented Feb 8 at 7:27
  • 1
    @ResistanceIsFutile Which tools are you using? That has not been my experience with either tools specifically designed for editing or using tools like ChatGPT and Bard for translation and editing tasks. Commented Feb 8 at 11:03
  • @ThomasOwens If you instruct AI to merely translate and fix grammatical errors, it may follow your instructions and usually you can get what you asked for, without it being too intrusive. But even then it may be recognizable as AI. However, the main problem is that people are commonly asking for improvement of their text and that results with rephrasing which is definitely leaks more AI traits and there are tools like Grammarly (which now use AI all over the place) and they also do more than simple fixes. Commented Feb 8 at 11:24
  • Now, for you and me, using any of those tools with care might not be a problem as we can recognize problematic parts, but for users that really need those because of their poor English or general writing skills, saying that such tools are allowed might get them in trouble as there is a high chance that such posts will be recognized as AI. Commented Feb 8 at 11:27
  • 1
    @ResistanceIsFutile I agree that there are risks if someone uses these tools, so saying that they are outright allowed doesn't make sense, so that's not what I'm asking for. I just think that we need clarity on integrating AI-generated content into human-generated content and making those lines a little more clear. Software Engineering has prohibited generative AI content, but paragraphs 2 and 3 both represent real issues that show how AI content can fit into communities. Commented Feb 8 at 11:42
  • I agree that we need clarity on what is allowed and when. I mean in some circumstances and on some sites, you would actually need to post part of some AI generated content as example or output and results you are having, obviously such usage should be allowed. But, having quoted AI in sense that person answering used AI to directly provide parts of the answer that we would otherwise expect to be written by person, even if that AI is quoted is definitely not something that sits right with me. Commented Feb 8 at 14:16
  • 1
    It seems Grammarly isn't just tweaking text nowadays. Commented Feb 11 at 2:44
  • @RebeccaJ.Stones Interesting. I knew they had AI tools for tone and other improvements, but not the full-up prompt-based generation. I just installed the plugin, since the free version has it. The improvements that take the full context of the text are nice and seem safe (in my limited testing), but the generative aspects are worrying. So it may depend on which aspects of the tool you are using. Commented Feb 11 at 11:14
2

— see revision 13 for changes that address this concern


Artificial intelligence (a.k.a. GPT, LLM, generative AI, genAI) tools can be used to generate content for $SiteName,

Is it usual to use the word "artificial intelligence" to refer to generative AI in English?

In Japan, artificial intelligence (AI) refers to a broader concept than generative AI, and it sometimes includes machine translation tools, search engines, or even compilers in some specific contexts. As one of the translators on Stack Overflow for Japanese, it is a bit hard to translate this sentence without an additional explanation, which I'd like to avoid for the clean translation.

So, how about changing wording a little for ease of translation? At least, GPT is not an alternative name of AI. While using the term "AI" is acceptable, could we possibly use more precise language? (If not using the word AI, we also should change the URL path /help/ai-policy, which could cause other complications....)

3
  • 3
    I think it makes sense to cover the broader AI landscape. If a new paradigm is invented tomorrow, we should by default cover that under the existing policy, rather than scramble again to figure out how to respond.
    – tripleee
    Commented Feb 9 at 7:15
  • 2
    Just for clarity, my suggestion does not intend to restrict a topic of the new help page to generative AI. The point I want to discuss is more precision in our wording. If the help page is solely for generative AI, then it should refers only generative AI instead of AI in general. If not, the page should describe generative AI just as an example.
    – nekketsuuu
    Commented Feb 12 at 0:49
  • 4
    Thank you for your feedback, @nekketsuuu — I have changed the language of the help center article drafts to refer to "generative artificial intelligence tools" to disambiguate.
    – JNat StaffMod
    Commented Feb 16 at 15:08
1

Most of what was written seemed fairly reasonable, but this seems strange...

Due to the ease of using artificial intelligence services, if a user wanted an answer from an artificial intelligence tool, it can be assumed that they would already have sought one.

Can it? The assumption is false for me; I'd be thrilled to get any useful answers to my questions. I don't care who/what it's written by.

But also...

  1. Sure, it's easy to get some answer from some genAI; it's not necessarily easy to get a useful answer to your precise question (that's literally the problem we're tackling). I also don't think genAI is anywhere near as widely used nor easily accessible as the text makes it sound. My impression is that most users are not even thinking of genAI when asking questions.

  2. Many Stack Exchange questions were asked before ChatGPT, etc., was even heard of. So the assumption cannot be applicable to such questions.

  3. While the OP might not "want" a genAI answer,

    • they also don't want us to downvote nor close their question too, yet we still do it;

    • OPs don't normally get to pick-and-choose who answers their questions, nor stipulate what tools answerers are allowed to use when writing an answer;

    • maybe the thousands of readers (other than the OP) simply want a useful answer to the question, regardless of how it's written;

    • surely what the OP actually wants is an answer to their question!

May I suggest something like...

Users who ask questions on $SiteName may have already sought answers elsewhere. Due to the ease of using artificial intelligence services, if a user wanted an answer from an artificial intelligence tool, it can be assumed that they would already have sought one, such as through artificial intelligence services. If they did not want an answer from an artificial intelligence tool, they can decide to ignore one that is marked as having been generated by it. Appropriate references help the reader decide whether or not they even want to read a given post, and flags the potential hazards involved in posts that have been written while using artificial intelligence tool.


Oh...

(i.e. optimization, security, etc.), often add excessive noise to their answers (i.e. explaining all details, no matter how relevant)

I think this confuses i.e. and e.g. Also I was taught that after i.e. and e.g. you're meant to add a comma. It seems like the first set of examples would be more relevant for sites where genAI is simply banned.


— see revision 13 for changes that address this concern

Sorry, one more thing: can you add examples of how e.g. ChatGPT might be acknowledged please?

7
  • 3
    I think you have missed the point. If someone wants AIG platforms to answer their question, they will ask on the AI platform and get answered on the AIG platform. Stack Exchange is a pointless third wheel to this, so the very fact it has been asked here at all is strong evidence that an AIG platform has not, or will not, or cannot provide the desired answer. Therefore, there is zero reason to think any question on Stack Exchange should be answered on Stack Exchange, by an AIG platform.
    – Nij
    Commented Feb 9 at 7:32
  • 3
    I don't object to your objections per se; but they don't convince me that the text is strange or unnecessary. The pressure should be on the presumtive AI answerer to find these (and maybe other) reasons to post their answer even though the policy carefully advises against posting trivial AI answers.
    – tripleee
    Commented Feb 9 at 7:32
  • 2
    Remember: in this thread, we're only talking about the sites where posts written with the help of AI are permissible. Useful contributions are, as always, welcomed and encouraged. We just want authors of posts that use AI help to acknowledge how it was used. Commented Feb 9 at 8:40
  • 2
    A good chunk of the AI posts are literally someone asking a question to chatgpt, and posting it, without reading it or understanding the question A lot of these people are doing the metaphorical equivilent of asking a question from here on say quora or the hyphen site, then taking those answers and cross posting them, which to me feels pretty abusive Commented Feb 9 at 9:11
  • 2
    If people could have just gone to the ai platform for answers instead of coming to SE, they could have just as easily googled/gone to a blog/read it from a book, etc and SE would still be unnecessary. This forum is needed so that people are spared the time spent on that research and they get an answer thats hopefully verified by a real person. There is a big difference between using ai as a tool to get to the answer and just copy/pasting whatever it gives out when asking it something
    – user13267
    Commented Feb 9 at 10:53
  • Maybe they could have just as easily done the research themselves. Or, more likely, that research is difficult and time-consuming, and a question on Stack Exchange is faster to receive an appropriate answerfrom an expert who knows it offhand and has the exaxt references on their bookmarks if not their desk. This is not comparable to a person asking on Stack Exchange for someone else to copy-paste it into ChatGPT or whatever AIG is coolest today, which would take longer than just s direct query to the AIG.
    – Nij
    Commented Feb 12 at 10:32
  • 1
    Knowing an answer "offhand and has the exaxt references on their bookmarks if not their desk." was and should never be a requirement for participating on the site. People are free to do their research and attempt to answer the question, and its nobody's business what tools they use for this research. Other users of the site should only be concerned with whether what actually gets posted on the site is useful or not,rather than obsess over what they think was used to create that post
    – user13267
    Commented Feb 12 at 13:21

You must log in to answer this question.

Not the answer you're looking for? Browse other questions tagged .