Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Computer vision
(section)
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
==Related fields== [[File:Intersection_over_Union_-_object_detection_bounding_boxes.jpg|thumb|[[Object detection]] in a photograph]] ===Solid-state physics=== [[Solid-state physics]] is another field that is closely related to computer vision. Most computer vision systems rely on [[image sensors]], which detect [[electromagnetic radiation]], which is typically in the form of either [[visible light|visible]], [[infrared light|infrared]] or [[ultraviolet light]]. The sensors are designed using [[quantum physics]]. The process by which light interacts with surfaces is explained using physics. Physics explains the behavior of [[optics]] which are a core part of most imaging systems. Sophisticated [[image sensors]] even require [[quantum mechanics]] to provide a complete understanding of the image formation process.<ref name="Szeliski2010" /> Also, various measurement problems in physics can be addressed using computer vision, for example, motion in fluids. ===Neurobiology=== {{multiple image | direction = horizontal | total_width = 400 | footer = | image1 = Simplified neural network training example.svg | alt1 = | caption1 = Simplified example of training a neural network in object detection: The network is trained by multiple images that are known to depict [[starfish]] and [[sea urchin]]s, which are correlated with "nodes" that represent visual [[Feature (computer vision)|features]]. The starfish match with a ringed texture and a star outline, whereas most sea urchins match with a striped texture and oval shape. However, the instance of a ring-textured sea urchin creates a weakly weighted association between them. | image2 = Simplified neural network example.svg | alt2 = | caption2 = Subsequent run of the network on an input image (left):<ref>{{cite book|author=Ferrie, C. |author2=Kaiser, S.|year=2019|title=Neural Networks for Babies|publisher=Sourcebooks|isbn=978-1492671206}}</ref> The network correctly detects the starfish. However, the weakly weighted association between ringed texture and sea urchin also confers a weak signal to the latter from one of two intermediate nodes. In addition, a shell that was not included in the training gives a weak signal for the oval shape, also resulting in a weak signal for the sea urchin output. These weak signals may result in a [[false positive]] result for sea urchin.<br>In reality, textures and outlines would not be represented by single nodes, but rather by associated weight patterns of multiple nodes.}} [[Neurobiology]] has greatly influenced the development of computer vision algorithms. Over the last century, there has been an extensive study of eyes, neurons, and brain structures devoted to the processing of visual stimuli in both humans and various animals. This has led to a coarse yet convoluted description of how natural vision systems operate in order to solve certain vision-related tasks. These results have led to a sub-field within computer vision where artificial systems are designed to mimic the processing and behavior of biological systems at different levels of complexity. Also, some of the learning-based methods developed within computer vision (''e.g.'' [[Artificial neural network|neural net]] and [[deep learning]] based image and feature analysis and classification) have their background in neurobiology. The [[Neocognitron]], a neural network developed in the 1970s by [[Kunihiko Fukushima]], is an early example of computer vision taking direct inspiration from neurobiology, specifically the [[Visual cortex#Primary visual cortex (V1)|primary visual cortex]]. Some strands of computer vision research are closely related to the study of [[biological vision]]βindeed, just as many strands of [[Artificial intelligence|AI]] research are closely tied with research into human intelligence and the use of stored knowledge to interpret, integrate, and utilize visual information. The field of biological vision studies and models the physiological processes behind visual perception in humans and other animals. Computer vision, on the other hand, develops and describes the algorithms implemented in software and hardware behind artificial vision systems. An interdisciplinary exchange between biological and computer vision has proven fruitful for both fields.<ref name="TextbookP1" /> ===Signal processing=== Yet another field related to computer vision is [[signal processing]]. Many methods for processing one-variable signals, typically temporal signals, can be extended in a natural way to the processing of two-variable signals or multi-variable signals in computer vision. However, because of the specific nature of images, there are many methods developed within computer vision that have no counterpart in the processing of one-variable signals. Together with the multi-dimensionality of the signal, this defines a subfield in signal processing as a part of computer vision. ===Robotic navigation=== [[Robot navigation]] sometimes deals with autonomous [[path planning]] or deliberation for robotic systems to [[robotic navigation|navigate through an environment]].<ref>Murray, Don, and Cullen Jennings. "[http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.65.6725&rep=rep1&type=pdf Stereo vision-based mapping and navigation for mobile robots] {{Webarchive|url=https://web.archive.org/web/20201031141636/http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.65.6725&rep=rep1&type=pdf |date=2020-10-31 }}." Proceedings of International Conference on Robotics and Automation. Vol. 2. IEEE, 1997.</ref> A detailed understanding of these environments is required to navigate through them. Information about the environment could be provided by a computer vision system, acting as a vision sensor and providing high-level information about the environment and the robot ===Visual computing=== {{excerpt|Visual computing}} ===Other fields=== Besides the above-mentioned views on computer vision, many of the related research topics can also be studied from a purely mathematical point of view. For example, many methods in computer vision are based on [[statistics]], [[Optimization (mathematics)|optimization]] or [[geometry]]. Finally, a significant part of the field is devoted to the implementation aspect of computer vision; how existing methods can be realized in various combinations of software and hardware, or how these methods can be modified in order to gain processing speed without losing too much performance. Computer vision is also used in fashion eCommerce, inventory management, patent search, furniture, and the beauty industry.<ref>{{cite web |last1=Andrade |first1=Norberto Almeida |title=Computational Vision and Business Intelligence in the Beauty Segment - An Analysis through Instagram |url=http://jmm-net.com/journals/jmm/Vol_7_No_2_December_2019/2.pdf |website=Journal of Marketing Management |publisher=American Research Institute for Policy Development |access-date=11 March 2024}}</ref> ===Distinctions=== The fields most closely related to computer vision are [[image processing]], [[image analysis]] and [[machine vision]]. There is a significant overlap in the range of techniques and applications that these cover. This implies that the basic techniques that are used and developed in these fields are similar, something which can be interpreted as there is only one field with different names. On the other hand, it appears to be necessary for research groups, scientific journals, conferences, and companies to present or market themselves as belonging specifically to one of these fields and, hence, various characterizations which distinguish each of the fields from the others have been presented. In image processing, the input and output are both images, whereas in computer vision, the input is an image or video, and the output could be an enhanced image, an analysis of the image's content, or even a system's behavior based on that analysis. [[Computer graphics]] produces image data from 3D models, and computer vision often produces 3D models from image data.<ref name="3DVAE">{{Cite book |doi=10.1109/CVPR.2017.269 |last1=Soltani |first1=A. A. |last2=Huang |first2=H. |last3=Wu |first3=J. |last4=Kulkarni |first4=T. D. |last5=Tenenbaum |first5=J. B. |title=2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) |chapter=Synthesizing 3D Shapes via Modeling Multi-view Depth Maps and Silhouettes with Deep Generative Networks |year=2017 |pages=1511β1519 |hdl=1721.1/126644 |isbn=978-1-5386-0457-1 |s2cid=31373273 |hdl-access=free }}</ref> There is also a trend towards a combination of the two disciplines, ''e.g.'', as explored in [[augmented reality]]. The following characterizations appear relevant but should not be taken as universally accepted: * [[Image processing]] and [[image analysis]] tend to focus on 2D images, how to transform one image to another, ''e.g.'', by pixel-wise operations such as contrast enhancement, local operations such as edge extraction or noise removal, or geometrical transformations such as rotating the image. This characterization implies that image processing/analysis neither requires assumptions nor produces interpretations about the image content. * Computer vision includes 3D analysis from 2D images. This analyzes the 3D scene projected onto one or several images, ''e.g.'', how to reconstruct structure or other information about the 3D scene from one or several images. Computer vision often relies on more or less complex assumptions about the scene depicted in an image. * [[Machine vision]] is the process of applying a range of technologies and methods to provide imaging-based automatic inspection, process control, and robot guidance<ref name="NASAarticle"/> in industrial applications.<ref name="TextbookP1"/> Machine vision tends to focus on applications, mainly in manufacturing, ''e.g.'', vision-based robots and systems for vision-based inspection, measurement, or picking (such as [[bin picking]]<ref>{{Cite web | url=https://www.robots.com/blog/viewing/the-future-of-automated-random-bin-picking | title=The Future of Automated Random Bin Picking | access-date=2018-01-10 | archive-date=2018-01-11 | archive-url=https://web.archive.org/web/20180111164947/https://www.robots.com/blog/viewing/the-future-of-automated-random-bin-picking | url-status=live }}</ref>). This implies that image sensor technologies and control theory often are integrated with the processing of image data to control a robot and that real-time processing is emphasized by means of efficient implementations in hardware and software. It also implies that external conditions such as lighting can be and are often more controlled in machine vision than they are in general computer vision, which can enable the use of different algorithms. * There is also a field called [[imaging science|imaging]] which primarily focuses on the process of producing images, but sometimes also deals with the processing and analysis of images. For example, [[medical imaging]] includes substantial work on the analysis of image data in medical applications. Progress in [[Convolutional neural network|convolutional neural networks]] (CNNs) has improved the accurate detection of disease in medical images, particularly in cardiology, pathology, dermatology, and radiology.<ref>{{Cite journal |last1=Esteva |first1=Andre |last2=Chou |first2=Katherine |last3=Yeung |first3=Serena |last4=Naik |first4=Nikhil |last5=Madani |first5=Ali |last6=Mottaghi |first6=Ali |last7=Liu |first7=Yun |last8=Topol |first8=Eric |last9=Dean |first9=Jeff |last10=Socher |first10=Richard |date=2021-01-08 |title=Deep learning-enabled medical computer vision |journal=npj Digital Medicine |volume=4 |issue=1 |page=5 |doi=10.1038/s41746-020-00376-2 |pmid=33420381 |issn=2398-6352|pmc=7794558 }}</ref> * Finally, [[pattern recognition]] is a field that uses various methods to extract information from signals in general, mainly based on statistical approaches and [[artificial neural networks]].<ref>{{Cite journal|last1=Chervyakov|first1=N. I.|last2=Lyakhov|first2=P. A.|last3=Deryabin|first3=M. A.|last4=Nagornov|first4=N. N.|last5=Valueva|first5=M. V.|last6=Valuev|first6=G. V.|year=2020|title=Residue Number System-Based Solution for Reducing the Hardware Cost of a Convolutional Neural Network|url=|journal=Neurocomputing|volume=407|pages=439β453|doi=10.1016/j.neucom.2020.04.018|s2cid=219470398|quote=Convolutional neural networks (CNNs) represent deep learning architectures that are currently used in a wide range of applications, including computer vision, speech recognition, identification of albuminous sequences in bioinformatics, production control, time series analysis in finance, and many others.}}</ref> A significant part of this field is devoted to applying these methods to image data. [[Photogrammetry]] also overlaps with computer vision, e.g., [[stereophotogrammetry]] vs. [[computer stereo vision]].
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)