Jump to content
Main menu
Main menu
move to sidebar
hide
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Special pages
Niidae Wiki
Search
Search
Appearance
Create account
Log in
Personal tools
Create account
Log in
Pages for logged out editors
learn more
Contributions
Talk
Editing
Overfitting
(section)
Page
Discussion
English
Read
Edit
View history
Tools
Tools
move to sidebar
hide
Actions
Read
Edit
View history
General
What links here
Related changes
Page information
Appearance
move to sidebar
hide
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
===Consequences=== {{multiple image | total_width = 300 | footer = Overfitted [[generative model]]s may produce outputs that are virtually identical to instances from their training set.<ref name=earthquake/> | image1 = Anne Graham Lotz (October 2008).jpg | alt1 = | caption1 = A photograph of [[Anne Graham Lotz]] included in the training set of [[Stable Diffusion]], a [[text-to-image model]] | image2 = Ann graham lotz stable diffusion.webp | alt2 = | caption2 = An image generated by Stable Diffusion using the prompt "Anne Graham Lotz" }} The most obvious consequence of overfitting is poor performance on the validation dataset. Other negative consequences include: * A function that is overfitted is likely to request more information about each item in the validation dataset than does the optimal function; gathering this additional unneeded data can be expensive or error-prone, especially if each individual piece of information must be gathered by human observation and manual data entry.<ref name=hawkins /> * A more complex, overfitted function is likely to be less portable than a simple one. At one extreme, a one-variable linear regression is so portable that, if necessary, it could even be done by hand. At the other extreme are models that can be reproduced only by exactly duplicating the original modeler's entire setup, making reuse or scientific reproduction difficult.<ref name=hawkins /> * It may be possible to reconstruct details of individual training instances from an overfitted machine learning model's training set. This may be undesirable if, for example, the training data includes sensitive [[personally identifiable information]] (PII). This phenomenon also presents problems in the area of [[artificial intelligence and copyright]], with the developers of some generative deep learning models such as [[Stable Diffusion]] and [[GitHub Copilot]] being sued for copyright infringement because these models have been found to be capable of reproducing certain copyrighted items from their training data.<ref name=earthquake>{{cite web |work=Ars Technica |last=Lee|first=Timothy B. |date=3 April 2023 |title=Stable Diffusion copyright lawsuits could be a legal earthquake for AI |url=https://arstechnica.com/tech-policy/2023/04/stable-diffusion-copyright-lawsuits-could-be-a-legal-earthquake-for-ai/ }}</ref><ref name="Verge copilot">{{Cite web |last=Vincent |first=James |date=2022-11-08 |title=The lawsuit that could rewrite the rules of AI copyright |url=https://www.theverge.com/2022/11/8/23446821/microsoft-openai-github-copilot-class-action-lawsuit-ai-copyright-violation-training-data |access-date=2022-12-07 |website=The Verge |language=en-US}}</ref>
Summary:
Please note that all contributions to Niidae Wiki may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see
Encyclopedia:Copyrights
for details).
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Search
Search
Editing
Overfitting
(section)
Add topic