Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Kismet (robot)
(section)
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
== Software system == Kismet's social intelligence software system, or [[synthetic nervous system]] (SNS), was designed with human models of intelligent behavior in mind. It contains six subsystems<ref>Breazeal, Cynthia. ''Designing Sociable Robots''. The MIT Press, 2002</ref> as follows. ===Low-level feature extraction system=== This system processes raw visual and auditory information from cameras and microphones. Kismet's '''vision system''' can perform eye detection, motion detection and, albeit controversial, skin-color detection. Whenever Kismet moves its head, it momentarily disables its motion detection system to avoid detecting self-motion. It also uses its stereo cameras to estimate the distance of an object in its visual field, for example to detect threats—large, close objects with a lot of movement.<ref name="Kismet, the robot">{{cite web|title=Kismet, the robot|url=http://www.ai.mit.edu/projects/sociable/baby-bits.html}}</ref> Kismet's '''audio system''' is mainly tuned towards identifying affect in [[infant-directed speech]]. In particular, it can detect five different types of affective speech: approval, prohibition, attention, comfort, and neutral. The affective intent classifier was created as follows. Low-level features such as pitch mean and energy (volume) variance were extracted from samples of recorded speech. The classes of affective intent were then modeled as a [[gaussian mixture model]] and trained with these samples using the [[expectation-maximization algorithm]]. Classification is done with multiple stages, first classifying an utterance into one of two general groups (e.g. soothing/neutral vs. prohibition/attention/approval) and then doing more detailed classification. This architecture significantly improved performance for hard-to-distinguish classes, like ''approval'' ("You're a clever robot") versus ''attention'' ("Hey Kismet, over here").<ref name="Kismet, the robot"/> ===Motivation system=== Dr. Breazeal figures her relations with the robot as 'something like an infant-caretaker interaction, where I'm the caretaker essentially, and the robot is like an infant'. The overview sets the human-robot relation within a frame of learning, with Dr. Breazeal providing the scaffolding for Kismet's development. It offers a demonstration of Kismet's capabilities, narrated as emotive facial expressions that communicate the robot's 'motivational state', Dr. Breazeal: "This one is anger (laugh) extreme anger, disgust, excitement, fear, this is happiness, this one is interest, this one is sadness, surprise, this one is tired, and this one is sleep."<ref>Suchman, Lucy. "Subject Objects." Feminist Theory. 2011, pg. 127</ref> At any given moment, Kismet can only be in one emotional state at a time. However, Breazeal states that Kismet is not conscious, so it does not have feelings.<ref>Breazeal, Cynthia. ''Designing Sociable Robots''. The MIT Press, 2002, pg. 112</ref> ===Motor system=== Kismet speaks a proto-language with a variety of phonemes, similar to a baby's babbling. It uses the [[DECtalk]] voice synthesizer, and changes pitch, timing, articulation, etc. to express various emotions. Intonation is used to vary between question and statement-like utterances. Lip synchronization was important for realism, and the developers used a strategy from animation:<ref>Madsen, R. ''Animated Film: Concepts, Methods, Uses''. Interland, New York, 1969</ref> "simplicity is the secret to successful lip animation." Thus, they did not try to imitate lip motions perfectly, but instead "create a visual shorthand that passes unchallenged by the viewer."
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)