Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Overfitting
(section)
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
===Consequences=== {{multiple image | total_width = 300 | footer = Overfitted [[generative model]]s may produce outputs that are virtually identical to instances from their training set.<ref name=earthquake/> | image1 = Anne Graham Lotz (October 2008).jpg | alt1 = | caption1 = A photograph of [[Anne Graham Lotz]] included in the training set of [[Stable Diffusion]], a [[text-to-image model]] | image2 = Ann graham lotz stable diffusion.webp | alt2 = | caption2 = An image generated by Stable Diffusion using the prompt "Anne Graham Lotz" }} The most obvious consequence of overfitting is poor performance on the validation dataset. Other negative consequences include: * A function that is overfitted is likely to request more information about each item in the validation dataset than does the optimal function; gathering this additional unneeded data can be expensive or error-prone, especially if each individual piece of information must be gathered by human observation and manual data entry.<ref name=hawkins /> * A more complex, overfitted function is likely to be less portable than a simple one. At one extreme, a one-variable linear regression is so portable that, if necessary, it could even be done by hand. At the other extreme are models that can be reproduced only by exactly duplicating the original modeler's entire setup, making reuse or scientific reproduction difficult.<ref name=hawkins /> * It may be possible to reconstruct details of individual training instances from an overfitted machine learning model's training set. This may be undesirable if, for example, the training data includes sensitive [[personally identifiable information]] (PII). This phenomenon also presents problems in the area of [[artificial intelligence and copyright]], with the developers of some generative deep learning models such as [[Stable Diffusion]] and [[GitHub Copilot]] being sued for copyright infringement because these models have been found to be capable of reproducing certain copyrighted items from their training data.<ref name=earthquake>{{cite web |work=Ars Technica |last=Lee|first=Timothy B. |date=3 April 2023 |title=Stable Diffusion copyright lawsuits could be a legal earthquake for AI |url=https://arstechnica.com/tech-policy/2023/04/stable-diffusion-copyright-lawsuits-could-be-a-legal-earthquake-for-ai/ }}</ref><ref name="Verge copilot">{{Cite web |last=Vincent |first=James |date=2022-11-08 |title=The lawsuit that could rewrite the rules of AI copyright |url=https://www.theverge.com/2022/11/8/23446821/microsoft-openai-github-copilot-class-action-lawsuit-ai-copyright-violation-training-data |access-date=2022-12-07 |website=The Verge |language=en-US}}</ref>
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)