Happiness, the AI way - PDF Free Download (2024)

Technology

Happiness, the AI way “BRIAN? How are you, Brian?” Thevoice is coming from a screen dominated by a vast blue cartoon eyeball, its pupil dilating in a way that makes it look both friendly and quizzical. Think HAL reimagined by Pixar. This is EmoSPARK, and it is looking for its owner. Its camera searches its field of view for a face and, settling on mine, asks again ifI am Brian. It sounds almost plaintive. –Feeling… boxed in– EmoSPARK’s brain is a 90-millimetre Bluetooth and WiFi-enabled cube. It senses its world yours, seeking always to through an internet connection, understand what makes you amicrophone, a webcam and your happy and unhappy. smartphone. Using these, the The “Brian” in question is Brian cube can respond to commands Fitzpatrick, a founding investor toplay any song in your digital in Emoshape, the company that library, make posts on Facebook makes EmoSPARK. He and the and check for your friends’ latest device’s inventor, Patrick Levy updates, stream a Netflix film, Rosenthal, compare EmoSPARK’s answer questions by pulling guiding principles to Isaac information from Wikipedia, Asimov’s laws of robotics. They and simply make conversation. are billing the cube as the world’s But its mission is more first “emotional AI”. complex: EmoSPARK, say its But EmoSPARK isn’t the first creators, is dedicated to your robotic agent designed to learn happiness. To fulfil that, it tries from our emotions. There’s Jibo totake your emotional pulse, the family robot and Pepper adapting its personality to suit the robot companion. Even

Feelings can sway robot choices, too Artificial intelligence works when the programmer has a specific goal in mind, such as collision avoidance. Butwhat about something more open-ended, such as foreseeing risk? This requires the human capacity to make judgements. One approach is to equip the machines with emotions such as fear, curiosity or frustration, says Massimiliano Versace at Boston University. Such emotions are an important aspect of our intelligence

20 | NewScientist | 16 May 2015

and decision-making, but are different from the social emotions now in vogue in AIs (see main story). These motivational emotions might be invisible to users of the AI, but more often than not, Versace says, the winning strategy “is the one that feels better”. He and his team have started working with NASA to design robot brains with emotional intelligence, to be used for exploring planetary surfaces.

Amazon’s Echo voice-activated controller might soon be able to recognise emotions. The drive to give artificial intelligence an emotional dimension is down to necessity, says Rana el Kaliouby, founder ofAffectiva, a Boston-based company that creates emotionsensing algorithms. As everything around us, from phones to fridges, gets connected to the internet, we need a way to temper machine logic with something more human. And when the user is immersed in a world that is as much computer as real life, a machine must learn some etiquette. For example, you shouldn’t come home from a funeral to find your AI itching to tell you about the latest Facebook cat videos. How can a machine be trained to understand emotions and act on them? When EmoSPARK’s webcam finds my face, a red box flashes briefly on screen to indicate it has identified a face that isn’t Brian’s. Behind the scenes, it is also looking for deeper details. EmoSPARK senses the user’s emotional state with the help of an algorithm that maps 80 facial points to determine, among other things, whether he or she is smiling, frowning in anger or sneering in disgust. EmoSPARK also analyses the user’s tone of voice, a long-established method of mood analysis. Having sensed these details, EmoSPARK uses them to mirror your emotions. First, it creates an emotional profile of its owner based on the combination of facial and voice input. At the end of each day, it sends this information to EmoShape,

Betsie Van Der Meer/Getty; left Emo Spark

Gadgets with emotional intelligence will soon be bonding with us to try to bring joy into our lives, finds Sally Adee

which sends back a newly tailored emotional profile for that particular device. Through this feedback loop, Fitzpatrick says, the cube’s personality changes ever so slightly every day.

Hard problems Rosalind Picard at the Massachusetts Institute of Technology is sceptical that this can produce an accurate emotional profile. Picard, who designs facial and vocal analysis software to help computers interpret emotion, and co-founded Affectiva with el Kaliouby, says there’s more to understanding moods than mapping points on the face. “What does it know aboutthe context? How much data is it trained on? How is it being taught the true feelings of the person? These are still hard

ONE PER CENT

like a joke it tells you, it won’t tell you that joke again,” he says. Until EmoSPARK has spent some time in people’s homes, we won’t know whether it can live up to its promise, or even whether having an AI trained on your emotional profile will make anyone feel happy. By now, however, 133 of EmoSPARK’s early crowdfunders have received their cubes and will act as beta testers. About 800 more should be available this month. Whether EmoSPARK succeeds or fails, AI with EQ is something we can expect to see much more

“We just can’t help projecting emotions on to anything from dolphins to Microsoft’s paper clip”

problems to solve.” The algorithm used by EmoSPARK isn’t necessarily all that sophisticated. Coaxing it to register a user’s smile requires a toothy grin in good lighting; realworld conditions, for most people, don’t live up to that. But maybe you don’t need a million-dollar algorithm. One aspect of creating “emotional” AI requires neither hardware nor software: it’s just a matter of exploiting what our brains do naturally. “We anthropomorphise everything,” says Eleanor Sandry at Curtin University in Perth, Australia. Humans project intent and emotions on to anything from dolphins to Microsoft’s paper clip. We can’t help ourselves. And EmoSPARK pulls out all the stops to put this tendency to work. To calibrate your cube, you undertake a ritual which ensures

of, says el Kaliouby. She believes all devices will one day have emotion processors, much as they now contain a GPS chip. This means every device will have its own proprietary algorithm for interpreting users’ emotions, and will reflect them back at the user in slightly different ways. If your –Do I detect a smile?– TV and your phone treat you a bit differently, that only adds to the that only one person can be illusion that you are surrounded emotionally bound to it. “Are you by a sentient cast of characters, the person I am to bond with?” is she says. its first question. Although it will Two weeks ago, Affectiva recognise other individuals in the released a mobile software same house or building, it only development kit which will creates the emotional profile for allow smartphone and tablet its owner. programmers to use its Affdex That doesn’t mean it can’t algorithm to assess emotions. interact with anyone else. When Some prototype applications someone who is not Brian taunts it, are already up and running. saying “I don’t like you”, EmoSPARK Chocolate firm Hershey’s is manifests its displeasure with a using Affdex to determine pulse of green light that shudders whether people smile at a candy through the cube. “It’s funny, I dispenser. If it detects a smile, the don’t like you that much either,” user gets a free chocolate sample. it responds. If EmoSPARK had Another is an art installation been complimented, it would that reads the facial expressions have glowed purple. of passers-by and composes Fitzpatrick says EmoSPARK can messages in real time on a wall to react to the user in more subtle cheer up the depressed and cheer ways, too, such as by withholding on the happy. “The idea that you information or trivia that it can measure emotion and act on regards as having displeased its it?” says el Kaliouby. “That’s owner previously. “If you don’t happened.” n

Tag it, smell it Graffiti artists, beware. Trains in Sydney, Australia, can now smell when you are up to no good. An undisclosed number have been fitted with electronic chemical sensors that can detect the vapours emitted by spray paint and permanent markers. When the sensors pick up a suspicious smell, live CCTV in the train sends images directly to security staff. So far, more than 30 people have been apprehended, say police.

“This may go down in history as the ‘it’s not our fault’ study” Internet researcher Christian Sandvig on Facebook’s paper in Science, which claims that individual choices – rather than its algorithms – create the “filter bubble” effect, which governs what a user sees and doesn’t see on social media

Virtual reality on sale It’s almost time for everyone to get immersive. Oculus Rift says its consumer virtual reality headset will go on sale in early 2016. Details, including the price, are scant. Oculus says the consumer version, based on recent prototypes, is lighter than the developers’ kits. It will also have an improved tracking system that will allow wearers to sit or stand while immersed in another world.

16 May 2015 | NewScientist | 21

Brendon Thorne/Getty

For more technology stories, visit newscientist.com/technology

Happiness, the AI way - PDF Free Download (2024)

References

Top Articles
Latest Posts
Recommended Articles
Article information

Author: Ms. Lucile Johns

Last Updated:

Views: 6252

Rating: 4 / 5 (41 voted)

Reviews: 88% of readers found this page helpful

Author information

Name: Ms. Lucile Johns

Birthday: 1999-11-16

Address: Suite 237 56046 Walsh Coves, West Enid, VT 46557

Phone: +59115435987187

Job: Education Supervisor

Hobby: Genealogy, Stone skipping, Skydiving, Nordic skating, Couponing, Coloring, Gardening

Introduction: My name is Ms. Lucile Johns, I am a successful, friendly, friendly, homely, adventurous, handsome, delightful person who loves writing and wants to share my knowledge and understanding with you.