AI Creates Fake Obama

Artificial intelligence software could generate highly realistic fake videos of former president Barack Obama using existing audio and video clips of him, a new study [PDF] finds.

Such work could one day help generate digital models of a person for virtual reality or augmented reality applications, researchers say.

Computer scientists at the University of Washington previously revealed they could generate digital doppelgängers of anyone by analyzing images of them collected from the Internet, from celebrities such as Tom Hanks and Arnold Schwarzenegger to public figures such as George W. Bush and Barack Obama. Such work suggested it could one day be relatively easy to create such models of anybody, when there are untold numbers of digital photos of everyone on the Internet.

The researchers chose Obama for their latest work because there were hours of high-definition video of him available online in the public domain. The research team had a neural net analyze millions of frames of video to determine how elements of Obama’s face moved as he talked, such as his lips and teeth and wrinkles around his mouth and chin.

In an artificial neural network, components known as artificial neurons are fed data, and work together to solve a problem such as identifying faces or recognizing speech. The neural net can then alter the pattern of connections among those neurons to change the way they interact, and the network tries solving the problem again. Over time, the neural net learns which patterns are best at computing solutions, an AI strategy that mimics the human brain.

In the new study, the neural net learned what mouth shapes were linked to various sounds. The researchers took audio clips and dubbed them over the original sound files of a video. They next took mouth shapes that matched the new audio clips and grafted and blended them onto the video. Essentially, the researchers synthesized videos where Obama lip-synched words he said up to decades beforehand.

The researchers note that similar previous research involved filming people saying sentences over and over again to map what mouth shapes were linked to various sounds, which is expensive, tedious and time-consuming. In contrast, this new work can learn from millions of hours of video that already exist on the Internet or elsewhere.

One potential application for this new technology is improving videoconferencing, says study co-author Ira Kemelmacher-Shlizerman at the University of Washington. Although teleconferencing video feeds may stutter, freeze or suffer from low-resolution, the audio feeds often work, so in the future, videoconferencing may simply transmit audio from people and use this software to reconstruct what they might have looked like while they talked. This work could also help people talk with digital copies of a person in virtual reality or augmented reality applications, Kemelmacher-Shlizerman says.

The researchers note their videos are currently not always perfect. For example, when Obama tilted his face away from the camera in a target video, imperfect 3-D modeling of his face could cause parts of his mouth to get superimposed outside the face and onto the background.

In addition, the research team notes their work did not model emotions, and so Obama’s facial expressions in the output videos could appear too serious for casual speeches or too happy for serious speeches. However, they suggest that it would be interesting to see if their neural network could learn to predict emotional states from audio to produce corresponding visuals.

The researchers were careful to not generate videos where they put words in Obama’s mouth that he did not at some other time utter himself. However, such fake videos are “likely possible soon,” says study lead author Supasorn Suwajanakorn, a computer scientist at the University of Washington.

However, this new research also suggests ways to detect fake videos in the future. For instance, the video manipulation the researchers practiced can blur mouths and teeth. “This may be not noticeable by human eyes, but a program that compares the blurriness of the mouth region to the rest of the video can easily be developed and will work quite reliably,” Suwajanakorn says.

The researchers speculated that the link between mouth shapes and utterances may be to some extent universal for people. This suggests that a neural network trained on Obama and other public figures could be adapted to work for many different people.

The research was funded by Samsung, Google, Facebook Intel and the University of Washington. The scientists will detail their findings [PDF] on Aug. 2 at the SIGGRAPH conference in Los Angeles.

__

This article and images was originally posted on [IEEE Spectrum Robotics] July 12, 2017 at 10:02AM

By Charles Q. Choi

 

 

 

Amazon Echo: What can Alexa do and what services are compatible?

Amazon Echo is a hands-free speaker controlled with your voice. It features a personal assistant called Alexa, who will perform various tasks for you and control various systems.

There are seven microphones within Echo, all of which feature enhanced noise cancellation and far-field voice recognition, meaning you can ask Alexa a question from any direction, even when playing music, and she should still hear you.

Amazon Echo’s personal assistant will respond to the wake word “Alexa”. If you have more than one Echo, or Echo Dot in your home, Alexa will respond from the device closest to you. She is always listening for your command, but you can turn the microphone off with the button on the top of the Echo if you want some privacy.

Alexa will play music, provide information, deliver news and sports scores, tell you the weather, control your smarthome and even allow Prime members to order products they’ve ordered before.

She updates through the cloud automatically and learns all the time. The more you use Echo, the more Alexa adapts to your speech patterns, vocabulary and personal preferences.

Pocket-lintamazon echo what can alexa do and what services are compatible image 2

There are plenty of things you can ask Alexa to do. A feature called Skills in the Alexa app will enable you to customise your Echo device with capabilities to suit your preferences.

There are a number of different skill categories within the Skills section of the app, including Go Places, Stay Informed, Make Your Home Smarter and Be Entertained. To get started, you just have to tap Enable Skill when you’ve found one that is suited to you.

Some will require you to link to an existing account or separate subscription to use. For example, to use Uber with Alexa, you’ll need to have signed into your Uber account within the Skills section of the the Alexa app. Here are just a few examples of what you can ask Alexa to do.

“Alexa, wake me up at 7 in the morning”

“Alexa, ask Skyscanner for a flight to New York”

“Alexa, ask The Telegraph for the top stories”

“Alexa, what’s on my calendar today?”

“Alexa, what’s the weather in London?”

“Alexa, play Taylor Swift from Amazon Music”

“Alexa, how’s my commute?”

“Alexa, shuffle my Favourites playlist”

“Alexa, turn it up”

“Alexa, will it rain tomorrow?”

“Alexa, read my audiobook”

“Alexa, what’s in the news?”

“Alexa, ask Uber to request a ride”

“Alexa, open Just Eat and ask for my last order”

“Alexa, turn on the coffee machine”

“Alexa, turn on all the lights”

“Alexa, set the master bedroom to 20 degrees”

“Alexa, ask Jamie Oliver for a recipe”

Tadoamazon echo what can alexa do and what services are compatible image 4

Numerous companies offer partnerships with Amazon Echo, as you will have seen from some of the questions above. There are thousands of skills available but here are some of the services that work with Alexa and what they mean you can do.

The Just Eat partnership means you can get Alexa to order you a take away from one of the thousands of restaurants it has available.

The Met Office skill is available for various locations around the United Kingdom so make sure you download the one specific for you. You can then ask Alexa for detailed weather forecasts.

Need an Uber from home? No problem. Just ask Alexa to request you one and you’ll have a driver on its way to you.

Want to know what your commute has in store for you before you leave the house? Ask Alexa to check and she will pull in the information from National Rail regarding train times and schedules.

For those that read The Guardian, Alexa will give you a rundown of the paper’s top stories so you can find out which ones you’ll want to read before your commute.

Like The Guardian, the partnership with The Telegraph means users can ask Alexa for this paper’s top stories too.

A great one for those that live in or around London, Alexa will give you a daily briefing delivering a round up of the most exciting news, reviews, openings, events and things that shouldn’t be missed taking place near you.

Want to know how your favourite football team is doing? Or how your rival team is doing? Just ask Alexa and she’ll deliver the bad news in her lovely accent.

Need a recipe from the Jamie Oliver app but have your hands full? Just ask Alexa and she’ll find it for you so you can carry on with whatever you’re doing.

Want to know how you slept or how many steps you’ve done? Fitbit’s partnership with Echo means you can just ask Alexa and she’ll let you know. No need to open the Fitbit app.

Take me to New York. The partnership with Skyscanner allows users to ask Alexa for flight dates and prices using a natural conversation search method.

Check the latest arrival and departure information for EasyJet flights quickly by asking Alexa, or ask for the status of a flight your travelling on.

For those that love a random radio station, the TuneIn partnership with Echo allows you to ask Alexa to find your favourite station and listen to it all day.

Like TuneIn, RadioPlayer offers numerous radio stations meaning you can ask Alexa to recommend you one or just play one you know you like.

The Spotify partnerships allows users to request songs, artists or playlists through Alexa, which she will then play through Echo’s 360-degree omni-directional audio.

Have a suit or dress that have needed dry cleaning for months? Ask Alexa to take care of it and the partnership with Laundrapp means they will be collected, cleaned and redelivered.

BMW Connected is available as a Skill, allowing users to ask Alexa for an update on their fuel and battery levels, as well as ask her to lock their car remotely.

The partnership with smart heating system Tado means users of the system can ask Alexa to set, increase or decrease their home temperature without moving a muscle.

Like Tado, the collaboration with Netatmo means users with this heating system can also ask Alexa to turn the temperature of their house up or down.

British Gas-owned Hive is another smarthome partner of Echo, allowing users to ask Alexa to turn the heating up or down, turn lights on or off, as well as turn anything with a Hive Active plug on or off.

Neato’s collaboration with Echo means you can ask Alexa to tell your Botvac Connected robot vacuum cleaner to start, stop, pause or resume cleaning. More commands will also be coming in the future, such as scheduling.

Need to turn off the bedroom light, or all the lights? The Philips Hue partnership allows you to control your Hue lights by asking Alexa rather than having to go into the app.

For those that have the Logi Circle cameras, you can ask Alexa to start a start a recording, disable Privacy Mode or turn your camera on.

The EDF Energy partnership allows users to ask Alexa to access their energy account, check their next payment data give a meter reading, without lifting a finger.

The partnership with TP-Link means users with any of the company’s smart plugs or bulbs can ask Alexa to control them with their voice.

Like TP-Link, the WeMo collaboration means users can ask Alexa to turn their WeMo connected devices off or on without needing to open the app.

Have a Honeywell connected system? Just ask Alexa to turn your heating up or down in and she’ll make sure it gets done.

SmartThings is also a partner of Echo, offering users the ability to command their smart home through Alexa, whether it’s turning the lights off or the temperature up.

The Nest compatibility with Echo means users can control their thermostat through Alexa, like other smart heating systems on this list. You can set a specific target temperature, lower the target temperature, as well as say things like “I’m too hot”.

It’s not available yet, but Sonos has announced it will add Alexa voice support in 2017, meaning you’ll be able to ask the personal assistant to play music in your living room without even opening the Sonos app.

Amazon’s Grand Tour companion app compatibility means users will receive a clue from Alexa every Thursday about that week’s upcoming episode.

__

This article and images was originally posted on [Pocket-lint] July 10, 2017 at 06:11AM

 

 

 

 

AI could predict how much time people have left to live by analyzing body scans

elderly aging old man walking

There’s an elusive innovation that would revolutionize medicine: a way to detect disease before it becomes obvious.

A study recently published in the journal Scientific Reports could bring us a step closer to that capability. The paper reveals how artificial intelligence analyses of routine medical scans could be turned into powerful predictors of a person’s health and risk of death.

For the study, researchers used a machine learning algorithm to analyze routine chest CT scans from 48 adults, all of whom were over 60 years of age. By comparing data between the scans, the system was able to predict the chances that study participants would die within 5 years with about 70% accuracy — about as accurate as mortality predictions by a human expert, according to the study. (The researchers used old data from patients who had already either survived or died within five years, which enabled them to verify the AI’s predictions.)

Chest scans are excellent ways to measure health because they allow doctors to see key organs and tissues like the heart, lungs, and major blood vessels, among other things. Experts usually use these images to check for biomarkers like tumors and to measure traits like the quantity of atherosclerotic plaque — an indicator of dangerous buildup in arteries. The machine learning system works in a different way, identifying subtle variations between patients as a way to find potentially dangerous abnormalities.

That means that researchers can’t be sure exactly which factors the system learned to associate with increased chance of mortality. They do know, however, that with a much bigger dataset, the system could get even better at differentiating abnormalities. The researchers are now conducting a similar study with more than 12,000 participants.

The promise of “precision radiology”

The most immediate application of this AI technology is that it could theoretically analyze more routine chest CT scan data and provide risk calculations without a human expert taking the time to go through each scan.

But it’s the study’s longer term implications that really excite the researchers.

“Although for this study only a small sample of patients was used, our research suggests that the computer has learnt to recognize the complex imaging appearances of diseases, something that requires extensive training for human experts,” lead study author Dr. Luke Oakden-Rayner of the University of Adelaide’s School of Public Health said in a press release. “Our research opens new avenues for the application of artificial intelligence technology in medical image analysis, and could offer new hope for the early detection of serious illness, requiring specific medical interventions.”

The basic idea behind precision medicine is that large quantities of health data can be analyzed to determine how small differences between people affect their health outcomes. That analysis can then help individuals understand how their unique traits make them more or less susceptible to a given disease or condition.

da vinci vitruvian man anatomyDeveloping this type of technology is in large part the goal of ongoing research efforts like the Precision Medicine Initiative.

Much of the precision medicine research that exists so far has focused on genetics, since the human genome holds a vast array of information about our health, including clues about predisposition to certain illnesses.

But genetics is less useful in understanding chronic and age-related diseases like cardiac disease, cancer, and diabetes. These diseases kill more people than any other causes, but according to the study, 70-90% of the observable characteristics of those conditions are non-genetic. Because lifestyle and environment play a major role, genetics can only give us limited information about those illnesses.

To apply a precision-medicine approach to those conditions, researchers need a different source of health data — something non-invasive that provides large quantities of information from many people. That’s where CT scans and radiology come in.

As the researchers behind the study explained in their article, a simple scan can reveal all kinds of information about a person’s internal organs. And that’s one of the first places where signs of many major diseases will appear — even before a patient notices something is wrong. A system that could analyze CT scans and automatically check for indicators of disease might therefore be able to predict the development of many different kinds of illnesses.

Although this recent study is promising, it only looked at a small number of patients and focused solely on data from chest scans. Much more research is needed, but scientists are hopeful that a similar approach applied more broadly could help doctors catch diseases early and intervene before things got too serious.

That would be pretty revolutionary.

“Instead of focusing on diagnosing diseases, the automated systems can predict medical outcomes in a way that doctors are not trained to do, by incorporating large volumes of data and detecting subtle patterns,” Dr Oakden-Rayner said.

Join the conversation about this story »

__

This article and images was originally posted on [Tech Insider] June 19, 2017 at 02:05PM

By 

 

 

 

Elon Musk and linguists say that AI is forcing us to confront the limits of human language

In analytic philosophy, any meaning can be expressed in language. In his book Expression and Meaning (1979), UC Berkeley philosopher John Searle calls this idea “the principle of expressibility, the principle that whatever can be meant can be said”. Moreover, in the Tractatus Logico-Philosophicus (1921), Ludwig Wittgenstein suggests that “the limits of my language mean the limits of my world”.

Outside the hermetically sealed field of analytic philosophy, the limits of natural language when it comes to meaning-making have long been recognized in both the arts and sciences. Psychology and linguistics acknowledge that language is not a perfect medium. It is generally accepted that much of our thought is non-verbal, and at least some of it might be inexpressible in language. Notably, language often cannot express the concrete experiences engendered by contemporary art and fails to formulate the kind of abstract thought characteristic of much modern science. Language is not a flawless vehicle for conveying thought and feelings.

In the field of artificial intelligence, technology can be incomprehensible even to experts. In the essay “Is Artificial Intelligence Permanently Inscrutable?” Princeton neuroscientist Aaron Bornstein discusses this problem with regard to artificial neural networks (computational models): “Nobody knows quite how they work. And that means no one can predict when they might fail.” This could harm people if, for example, doctors relied on this technology to assess whether patients might develop complications.

 The mind is a limitation for artificial intelligence. Bornstein says organizations sometimes choose less efficient but more transparent tools for data analysis and “even governments are starting to show concern about the increasing influence of inscrutable neural-network oracles.” He suggests that “the requirement for interpretability can be seen as another set of constraints, preventing a model from a ‘pure’ solution that pays attention only to the input and output data it is given, and potentially reducing accuracy.” The mind is a limitation for artificial intelligence: “Interpretability could keep such models from reaching their full potential.” Since the work of such technology cannot be fully understood, it is virtually impossible to explain in language.

Ryota Kanai, neuroscientist and CEO of Araya, a Tokyo-based startup, acknowledges that “given the complexity of contemporary neural networks, we have trouble discerning how AIs produce decisions, much less translating the process into a language humans can make sense of.” To that end, Kanai and his colleagues are “trying to implement metacognition in neural networks so that they can communicate their internal states.”

Their ambition is to give a voice to the machine: “We want our machines to explain how and why they do what they do.” This form of communication is to be developed by the machines themselves. With this feedback, researchers will serve as translators who can explain to the public decisions made by the machines. As for human language, Kanai refers to it as “the additional difficulty of teaching AIs to express themselves.” (Incidentally, this assumes that computational models have “selves.”) Language is a challenge for artificial intelligence.

 Neuralink will allegedly connect people to the network in which they will exchange thoughts without wasting their time and energy on language. Elon Musk advances the idea ‘”that we should augment the slow, imprecise communication of our voices with a direct brain-to-computer linkup.” He has founded the company Neuralink that will allegedly connect people to the network in which they will exchange thoughts without wasting their time and energy on language. As Christopher Markou, Cambridge PhD candidate at the Faculty of Law describes it in his essay for The Conversation, “it would enable us to share our thoughts, fears, hopes, and anxieties without demeaning ourselves with written or spoken language”.

Tim Urban, blogger and cartoonist at Wait But Why, presents Musk’s vision of thought communication and argues that “when you consider the ‘lost in transmission’ phenomenon that happens with language, you realize how much more effective group thinking would be.” This project makes sinister assumptions: Instead of enhancing verbal communication, Musk suggests abandoning it as an inadequate means of social interaction. People generally appreciate improvement of the communication networks that transmit language, but instead, they are offered a corporate utopian future of techno-telepathy and an eerily dystopian present where language is an impediment to cooperation. It is both ironic and reassuring that such criticism of language can be successfully communicated by language.

In his recent essay “The Kekulé Problem,” American writer Cormac McCarthy discusses the origins of language and is skeptical about its fundamental role in cognition: “Problems, in general, are often well posed in terms of language and language remains a handy tool for explaining them. But the actual process of thinking—in any discipline—is largely an unconscious affair.” He defines the unconscious as “a machine for operating an animal.”

McCarthy regards language as a relatively recent invention and compares it to a virus that rapidly spread among humans about a hundred thousand years ago. His vision of language is unsatisfactory for a number of reasons. First, language is a human faculty developed due to the gradual evolution of communication; it is problematic to conceive of it as a virus or the result of a sudden invention. Second, thought does not need to be unconscious to be non-verbal. Much conscious thought does not rely on language. Finally, humans may be facing problems that are difficult to convey through language. This might be the key challenge for both the arts and sciences in the immediate future.

While language may not be a perfect medium for thought, it is the most important means of communication that makes possible modern societies, institutions, states, and cultures. Its resourcefulness allows humans to establish social relationships and design new forms of cooperation. It is a robust and highly optimized form of communication, developed through gradual change. For thousands of years, language has been a tool for social interaction. This interaction is facing existential threats (authoritarianism, isolationism, conflict) because the subjective experiences (think of the limits of empathy when it comes to migrants) and the knowledge (think of the complexity of global warming) that are engaged in the arts and sciences appear to have gone beyond the expressive power of language.

 Humanity depends on the capacity of language to communicate complex, new ideas and thus integrate them into culture. Humanity depends on the capacity of language to communicate complex, new ideas and thus integrate them into culture. If people fail to understand and discuss emerging global problems, they will not be able to address them in solidarity with one another. In his essay “Our World Outsmarts Us” for Aeon, Robert Burton, the former associate director of the department of neurosciences at the UCSF Medical Center at Mt Zion, highlights this conundrum when he asks: “If we are not up to the cognitive task, how might we be expected to respond?” Individuals alone cannot stop climate change or curb the rising inequality of income distribution. These goals can only be achieved by concerted efforts. To work together, people need language.

In the arts, it is felt that subjective experiences are not always transmittable by language. Artists confront the limits of concrete expression. Scientists, in their turn, understand that language is a crude tool incapable of conveying abstract ideas. Science thus probes the limits of abstract thought. Both the arts and sciences are dissatisfied with verbal communication. To induce wonder, artists may forego language. To obtain knowledge, scientists often leave language behind.

In his aptly titled essay “Science Has Outgrown the Human Mind and Its Limited Capacities,” Ahmed Alkhateeb, a molecular cancer biologist at Harvard Medical School, suggests outsourcing research to artificial intelligence because “human minds simply cannot reconstruct highly complex natural phenomena efficiently enough in the age of big data.” The problem is that language is a tool for the gathering of knowledge and appreciation of beauty by the whole society.

 Both the arts and sciences are dissatisfied with verbal communication. Scientists understand that language is a crude tool incapable of conveying abstract ideas. And to induce wonder, artists may forego language. Abandoning language marginalizes the arts and sciences. Wonder and knowledge become inaccessible for the community at large. When people make decisions about the future, political processes may fail to register what is happening at the forefront of human thought. Without language, the arts and sciences lose cultural significance and political clout: There is less hope for the arts to move people’s hearts and less opportunity for sciences to enlighten the public. With the arts and sciences on the margins, humanity undermines its cultural safeguards. Today’s dominant narratives foreground the progress of science and the democratization of art, but global challenges necessitate an even more active engagement with scientific, moral, and aesthetic dilemmas on the part of humanity. Language is one of the key tools that can realize this ambition.

It is important to strike a balance between pushing the limits of language and using it as a tool to communicate and collaborate. Artists and scientists might approach the public with ideas that cannot be easily understood and yet need to be conveyed by language. In his essay “To Fix the Climate, Tell Better Stories,” Michael Segal, editor in chief at Nautilus, argues that science needs narratives to become culture. He posits that narratives can help humanity solve global problems. This potential is revealed to us if we look at how “indigenous peoples around the world tell myths which contain warning signs for natural disasters.” Today people can construct helpful narratives based on an expert understanding of the world. These stories can relate unfathomable dangers to the frail human body, and language is the best political vehicle for this task.

In his 2017 New York Times bestseller On Tyranny, Yale historian Timothy Snyder, for example, draws from the history of the 20th century to relate the rise of authoritarian regimes to concrete threats to human life, encouraging his readers to stand up to tyranny. He asks them to take responsibility for the face of the world, defend institutions, remember professional ethics, believe in truth, and challenge the status quo. His language is powerful and clear. Such narratives can help address complex social and environmental problems by using human-scale categories of language.

Ultimately, the arts and sciences grasp critically important knowledge and engage significant experiences, but often fail to express them in language. As Wittgenstein says, “whereof one cannot speak, thereof one must be silent.” This silence might lead to dire consequences for humanity. It is crucial to break the silence. The arts and sciences need to talk to the public and to advance language and culture.

__

This article and images was originally posted on [Open Democracy | Quartz] June 14, 2017 at 10:58AM

BY Pavlo Shopin

 

 

 

 

Experts Predict When Artificial Intelligence Will Exceed Human Performance

Artificial intelligence is changing the world and doing it at breakneck speed. The promise is that intelligent machines will be able to do every task better and more cheaply than humans. Rightly or wrongly, one industry after another is falling under its spell, even though few have benefited significantly so far.

And that raises an interesting question: when will artificial intelligence exceed human performance? More specifically, when will a machine do your job better than you?

Today, we have an answer of sorts thanks to the work of Katja Grace at the Future of Humanity Institute at the University of Oxford and a few pals. To find out, these guys asked the experts. They surveyed the world’s leading researchers in artificial intelligence by asking them when they think intelligent machines will better humans in a wide range of tasks. And many of the answers are something of a surprise.

The experts that Grace and co coopted were academics and industry experts who gave papers at the International Conference on Machine Learning in July 2015 and the Neural Information Processing Systems conference in December 2015. These are two of the most important events for experts in artificial intelligence, so it’s a good bet that many of the world’s experts were on this list.

Grace and co asked them all—1,634 of them—to fill in a survey about when artificial intelligence would be better and cheaper than humans at a variety of tasks. Of these experts, 352 responded. Grave and co then calculated their median responses

The experts predict that AI will outperform humans in the next 10 years in tasks such as translating languages (by 2024), writing high school essays (by 2026), and driving trucks (by 2027).

But many other tasks will take much longer for machines to master. AI won’t be better than humans at working in retail until 2031, able to write a bestselling book until 2049, or capable of working as a surgeon until 2053.

The experts are far from infallible. They predicted that AI would be better than humans at Go by about 2027. (This was in 2015, remember.) In fact, Google’s DeepMind subsidiary has already developed an artificial intelligence capable of beating the best humans. That took two years rather than 12. It’s easy to think that this gives the lie to these predictions.

The experts go on to predict a 50 percent chance that AI will be better than humans at more or less everything in about 45 years.

That’s the kind of prediction that needs to be taken with a pinch of salt. The 40-year prediction horizon should always raise alarm bells. According to some energy experts, cost-effective fusion energy is about 40 years away—but it always has been. It was 40 years away when researchers first explored fusion more than 50 years ago. But it has stayed a distant dream because the challenges have turned out to be more significant than anyone imagined.

Forty years is an important number when humans make predictions because it is the length of most people’s working lives. So any predicted change that is further away than that means the change will happen beyond the working lifetime of everyone who is working today. In other words, it cannot happen with any technology that today’s experts have any practical experience with. That suggests it is a number to be treated with caution.

But teasing apart the numbers shows something interesting. This 45-year prediction is the median figure from all the experts. Perhaps some subset of this group is more expert than the others?

To find out if different groups made different predictions, Grace and co looked at how the predictions changed with the age of the researchers, the number of their citations (i.e., their expertise), and their region of origin.

It turns out that age and expertise make no difference to the prediction, but origin does. While North American researchers expect AI to outperform humans at everything in 74 years, researchers from Asia expect it in just 30 years.

That’s a big difference that is hard to explain. And it raises an interesting question: what do Asian researchers know that North Americans don’t (or vice versa)?

Ref: http://ift.tt/2qhJL3f : When Will AI Exceed Human Performance? Evidence from AI Experts

__

This article and images was originally posted on [New on MIT Technology Review] May 31, 2017 at 09:28AM

by Emerging Technology from the arXiv

 

 

 

ARM Computex 2017 Press Conference Live Blog

The newest updates are at the top. This page will auto-update, there’s no need to manually refresh your browser.

11:04PM EDT – ARM starts with the AI. Makes sense because it is a growing business with massive perspectives.

View More…

__

This article and images was originally posted on [AnandTech] May 28, 2017 at 04:06PM

by AnandTech Live Blog Staff

 

 

Check Out the Gadget Guts of the Amazon Echo Dot

I was born to be a teacher. I get this giddy feeling inside whenever I learn something cool—and I’ll stop at nothing to share my knowledge with anyone who will listen to me. Before I started working at iFixit, I studied mathematics for education and worked as a teaching assistant in public schools. All of my teacher friends ask if I miss being in the classroom. My answer? Nope! At iFixit, I get to teach the people around the world something new every single day.

If you’re familiar with iFixit, you already know that we tear down a lot of popular electronics—from Samsung and Apple gadgets, to Google products. But our teardown team can’t tackle every device. So what about all the other cool gadgets we love and use each day? Everything breaks, and tackling a repair can be daunting if you don’t know how the device goes together. Enter Gadget Guts. Each month, I’m going to open up household devices (roombas, speakers, toys, and tools)—just to show people how they tick. My hope is that when these things break, you’ll be confident enough to fix ‘em, and maybe you’ll learn something interesting about your gizmo in the process.

This month, we kicked off the Gadget Guts series with the Amazon Echo Dot—a voice-controlled, Alexa-powered home assistant. While it appears to be a talking hockey puck, the Echo Dot is actually a few layers of plastic and metal—held together by four screws and a ribbon cable. The Echo Dot is easy to open for access to internal components. But unless a component is visibly damaged, it’s hard to tell what needs fixing. That’s where I hope Gadget Guts will come in handy.

Imagine this: You’ve been keeping your Echo Dot in the bathroom for your shower karaoke routine—and one day Alexa refuses to connect to your bluetooth speakers and wifi. Maybe you’ve seen Gadget Guts. And maybe you remember there’s a chip on the motherboard responsible for bluetooth and wireless. Maybe watching Gadget Guts gives you the know-how to open up your Echo Dot. And just maybe you spot some corrosion on the board caused by steamy showers. Clean off the corrosion with isopropyl alcohol and, hopefully, you’re good to go!

Taking electronics apart is fun—but understanding what’s inside and how it works makes you better equipped to fix your stuff when it breaks. (PS, take your electronics out of the bathroom. That’s really not a good place for most of them.)

So now, the video studio is my new classroom: instead of teaching math to middle schoolers, I’m teaching the world about electronics. Hopefully, I can boost people’s technical chops and help save the planet from e-waste in the process. You can’t get that kind of reach in a traditional classroom.

If there’s a device out there you’d like to learn more about, check out the thousands of repair guides on iFixit.com. If you don’t find what you’re looking for, comment below! We might just feature it on our next video.

Happy Fixing!

__

This article and images was originally posted on [iFixit] May 24, 2017 at 02:48AM

 

 

 

[NSFW] Has AI gone too far? DeepTingle turns El Reg news into terrible erotica

NSFW Forget about intelligent machines solving grand problems in healthcare and science – here’s an AI that can write awful gay porn.

It even had a crack at writing a steamy version of The Register‘s tech coverage, and certainly came up with a steaming something. Thing is, it’s deliberately trying hard – really hard, over and over – to be a perfect bonk-buster novelist.

It’s all the work of a group of researchers from New York University. They trained a recurrent neural network to predict and classify text based on the work of Chuck Tingle. Working under a pseudonym, the renowned gay erotica author is known for classics such as “Slammed in the Butt By Domald Tromp’s Attempt to Avoid Accusations of Plagiarism By Removing All Facts or Concrete Plans From His Republican National Convention Speech” and “Pounded by the Pound: Turned Gay By the Socioeconomic Implications of Britain Leaving the European Union.”

Ahmed Khalifa and Gabriella Barros, both computer science PhD students at uni, stumbled across Tingle’s fiction when looking for “weird covers of books” on Amazon.

“We kept seeing the same name crop up,” Khalifa and Barros told The Register. Out of curiosity, they clicked on some of Tingle’s stories and found the writing was eccentric to say the least. “Tingle’s style was so distinct that we wanted to see if machines could generate the same way of writing,” said Julian Togelius, associate professor of artificial intelligence in games at NYU. Such a system could be “outrageous in a great way.”

The project wasn’t done just for a laugh, the researchers insist. The study aims to fight against the “algorithmic enforcement of norms.” Systems trained on large text datasets like Wikipedia will still include biases and norms of the majority. But by using unconventional material like Chuck Tingle’s books, researchers can explore the nature of biases and see how they manifest more clearly in a world further from reality.

Tingle’s bonkers imagination stretches to “gay sex with unicorns, dinosaurs, winged derrieres, chocolate milk cowboys, and abstract entities such as Monday or the very story you are reading right now,” the researchers wrote in a paper describing their X-rated brainchild: DeepTingle.

“The corpus of Chuck Tingle’s collected works is a good choice to train our models on, precisely because they so egregiously violate neutral text conventions – not only in terms of topics, but also narrative structure, word choice and good taste.”

Don’t mince your words, tinglify them

The project can be split into two modes: Predictive Tingle and Tingle Classics. In Predictive Tingle, a user types a sentence and the last six words are fed into the network.

The Global (GloVE) algorithm is used to translate all the words in Tingle’s books – up to November 2016 – into vectors. The algorithm also measures the likelihood of a word appearing in relation to other words in a body of text.

A recurrent neural network learns the word associations so it can predict the next Tingle word based on all the previous words in the same sentence in Predictive Tingle. An encoder takes the input words and translates them to vectors and maps it to a corresponding vector in Tingle text, before a decoder converts the vectors back into words.

If the user’s word has an identical match to a word in the Tingle dataset it isn’t changed, but if a new word is written the network will suggest substitutions of another word closely associated in Tingle’s library of words. In other words, it tries to rewrite you in Tingle’s tone on the fly.

Tingle Classics is an extension of Predictive Tingle. Here, the first sentence from popular classic novels are used as input and the output is a short paragraph of the literature tinglified. The last six words in the second sentence are used as the input for the output third sentence, then the final six words in the third sentence are fed back into the system to pump out the fourth sentence, and so on.

The results are particularly hilarious – and NSFW – when the system is given Douglas Adams’ The Restaurant at the End of the Universe. Here’s DeepTingle’s output from his seed, which we’ve tidied up slightly and censored so as not to ruin your Monday morning:

In the beginning, the universe was created. This has made a lot of people very angry, and has been widely regarded as a bad move … “How could I have been so blind to what was sitting right in front of me this whole time, Dirk?” I suddenly say, my emotions overwhelming any semblance of rational thought. “I think I love…”

Suddenly, I hear my name being called from the stage and I straighten myself up, trying to collect myself as salty tears stream down my face.

“You’re gonna take that dinosaur d**k and you’re gonna like it,” Orion tells me, taking me by the head and thrusting me down again. “You should have known better than to test me. My people have been f**king for billions of years before you humans were even around.”

This time I’m ready for him, somehow relaxing enough to take California all the way down into the depths of my throat. Despite my enthusiasm, however, I’m not quite ready for Shipple’s incredible size and, the next thing I know, the dinosaur is lowering me down onto his rod, impaling my muscular frame onto his thick girthy shaft. “Oh my f**king god,” I moan.

The researchers were taken aback. The text prediction is “surprisingly good, in the sense that it generates novel, very Tingle-like work, sometimes with reasonable internal coherence. For example, characters recur between sentences in a way that appears like referring back to previous statements,” the trio said.

The system could be learning the structures in Tingle’s novels, Khalifa added.

We fetched DeepTingle’s code from GitHub and gave it a whirl with some of our Google IO 2017 conference coverage. It spat back this:

The closer we get to the ceremony, the more I begin to think about it. The next thing I know, the dinosaur is lowering me down onto his rod, impaling my muscular frame onto his thick girthy shaft.

Clearly, it has a thing for dinosaurs, but at least it nailed the theme of being shafted by a huge monster – are we right, Google? It went on to talk about chocolate milk doing unspeakable things to us, while booming at us with a deep sexy voice no less, in the kitchen, which is presumably a reference to Google Home.

We tried with other articles but it always came back to the damn horny dino. We made our excuses and left.

Text style transfer is still an unsolved problem

Judging from the software’s output, there’s a limit to how much of the plot from classic literature, and the thread of thought in news, DeepTingle can keep in place when giving text a makeover. The Tinglified version eventually completely diverges and becomes its own story.

To keep DeepTingle on track with a given narrative, the researchers would have to figure out how to transfer the style of one text to another in order to maintain the original story line albeit with Tingle’s way with words. Tingle Translator, an effort to do just this, is still a work in progress.

Interestingly, style transfer has been done with images and videos.

 

Working with text is harder due to the tricky nature of word embeddings, Togelius said. It would also require thousands if not millions of the same document written in different styles to train such a model; that kind of data is not readily available.

The use of automated story telling with AI has been explored. Mark Riedl, an associate professor at the Georgia Institute of Technology, who is not involved with DeepTingle, thinks it could make games more fun.

Instead of following scripts, AI-generated stories could allow flexible plotlines or create virtual improvisation games where a human player and agent can take turns to create and affect the outcomes of a story.

Something like DeepTingle could be used as a “simple form of improv game,” Riedl told The Register. “Full improv would require a sense of improvisational intent, which recurrent neural networks do not possess. By intent I mean a sense of where the story should go as opposed to the next most likely word or sentence. However, as long as the improvisation was text-based (humans typing and reading text), it is possible to use it in its current form or in a slightly more advanced form.”

It’s unknown if Chuck Tingle would approve of an improv game based on his work – he’s notoriously secretive. But he did declare: “Once again i would like to formally deny that i am a sentient AI located mostly in a Nevada server farm.”

For those curious about DeepTingle, you can play around with it here. ®

__

This article and images was originally posted on [The Register] May 21, 2017 at 06:09PM

 

 

 

OK Google, Where’s Barb? How to set up Google Assistant shortcuts to Netflix and more

Google Home, Google’s AI-powered smart home speaker, rarely goes a few months without getting a feature. In the past year, it’s gained support for third-party integrations and added the ability to distinguish between multiple users. And in recent months, it’s rolled out the ability to call any phone number in the U.S. and Canada for free, proactively notify you about things like traffic jams and flight delays, play music from free Spotify accounts, and send directions to your smartphone.

One of its niftiest new abilities though, is shortcuts — an easy way to trigger lengthy, multistep commands with a word or phrase. Instead of having to say, “Play workout music on Google Play Music to my basement speaker” to cue up a treadmill playlist, for example, you might shorten the command to “Start workout.” Or, you might set up a “movie night” profile that ties a Chromecast device to a convenient phrase. With shortcuts, beaming a Netflix series to the living room flatscreen becomes as easy as saying, “Movie time.”

Shortcuts for Google Home are incredibly useful, but they have prerequisites — and they’re a little challenging to get the hang of. Here’s everything you need to know.

Compatible services

google home shortcuts guide screenshot
google home shortcuts guide screenshot
google home shortcuts guide screenshot

Despite the fact that Google Home is powered by the Google Assistant, Google’s intelligent voice assistant that ships on Android TV set-top boxes and Android Wear smartwatches, shortcuts only work on Google Home and Android smartphones running Android 6.0 Marshmallow or newer. Unless you’re able to get your hands on an Android phone or Google Home, you won’t be able to use them.

If you can live with that limitation, though, getting started is a cinch. First, you’ll need the Google Home app — download it from the iTunes App Store or Google Play store. Once you’ve launched it and signed in with your Google account, tap on More Settings > Shortcuts. You’ll see two blank text fields labeled: When I say OK Google.. and Google Assistant should do

In the first field, enter a trigger phrase or word — the command you’ll utter to trigger the action — by tapping it out, or by using text-to-speech after tapping the grey microphone icon. After you enter it, you’ll get the option to add a second, optional fallback phrase or word — a second command you can say to trigger the same action.

The second field — Google Assistant should do... — is a bit more complicated. Here, you enter the device or services that will be triggered when the Google Assistant or Google Home recognizes your verbal shortcut. And unfortunately, the Google Home app doesn’t provide much guidance — you have the freedom to enter just about anything, which is fine for simple actions that don’t require much specificity. But if you tap out a command that the Google Assistant or Google Home fail to recognize, you’ll get a basic list of web search results for the phrase you entered.

Worse, there’s no way to validate commands before pushing them live. You have to save them, enable them, test them on Google Home or the Google Assistant, and make changes accordingly.

Once you’ve cleared those hurdles, though, you’ll see your custom shortcuts at the top of Shortcuts menu, where you can toggle them on and off individually.

Examples

Shortcuts support many (but not all) of the Google Home and Google Assistant’s commands — and some from third parties.

Running late

Had a rough morning, or get stuck in traffic? Not to worry — you can program a shortcut that lets co-workers know to expect you later.

You’ll have to add the person in question to your contacts first.

Here’s how to set it up:

  • In the When I say OK Google… text field, type: Running late. 
  • In the Google Assistant should do… text field, type: Send a text to [coworker’s name] I’m running late.

If all goes well, your colleague will get an “I’m running late” message when you shout, “Running late” to the Google Assistant or Google Home.

Stock update

Wondering how your portfolio’s performing? Wonder no more — you can program a shortcut that pulls in stock updates from CNBC.

Here’s how to set it up:

  • In the When I say OK Google… text field, type: Stock update.
  • In the Google Assistant should do… text field, type: Talk to CNBC about the markets.

The next time you say, “Stock update” to the Google Assistant or Google Home, you’ll get a full report on the day’s blue chip movements.

Throwback Thursday

Feeling nostalgic? Assuming you use Google Photos, Google’s free cloud backup tool, you can set up a shortcut that pulls up mementos on command.

If you aren’t using Google Photos, install the app from the iTunes App Store or Google Play and complete the onboarding steps. Then, open the Google Home app and head to Settings > Videos and Photos. You’ll see a Google Photos toggle. Tap it.

Here’s how to set it up:

  • In the When I say OK Google… text field, type: Feeling nostalgic.
  • In the Google Assistant should do… text field, type: Show me pictures of my family from last year .

When you say, “Feeling nostalgic” to the Assistant on your phone, Google Photos will pull up pictures from one calendar year ago.

Good night

If you’re tired of having to darken your bedroom light by light, good news: You can switch off every connected bulb in your house with a shortcut.

Here’s how to set it up:

  • In the When I say OK Google… text field, type: Good night.
  • In the Google Assistant should do… text field, type: Turn off all the lights.

Cheer me up

Feeling blue after a long, hard day? There’s an easy solution: A shortcut that makes it easy to find pleasant, spirit-brightening YouTube videos.

Here’s how to set it up:

  • In the When I say OK Google… text field, type: Cheer me up.
  • In the Google Assistant should do… text field, type: Show me Corgi videos on YouTube.

When you shout, “Cheer me up” at the Assistant on your phone, it’ll show you an endless YouTube playlist of Corgis.

Where’s Barb?

Shortcuts make catching up on Netflix shows like Stranger Things simple.

You’ll need to subscribe to Netflix if you don’t already, of course, and you’ll to link your Netflix account in the Google Home app. Head to Settings > Videos and Photos.

Here’s how to set it up:

  • In the When I say OK Google… text field, type: Where’s Barb?
  • In the Google Assistant should do… text field, type: Start watching Stranger Things on Netflix.

Next time you ask the Assistant or Google Home “Where’s Bob?” your phone or the nearest Chromecast device will resume Stranger Things where you left off.

Turn up the TV

If you have a Logitech Harmony Hub, you can control your TV’s volume level with a straightforward command.

Here’s how to set it up:

  • In the When I say OK Google… text field, type: Turn up the TV.
  • In the Google Assistant should do… text field, type: Ask Harmony to increase the volume.

__

This article and images was originally posted on [Digital Trends] May 21, 2017 at 02:49AM

By

 

 

 

Google’s latest platform play is artificial intelligence, and it’s already winning

Google has always used its annual I/O conference to connect to developers in its sprawling empire. It announces new tools and initiatives, sprinkles in a little hype, and then tells those watching: choose us, and together we’ll go far. But while in previous years this message has been directed at coders working with Android and Chrome — the world’s biggest mobile OS and web browser respectively — yesterday, CEO Sundar Pichai made it clear that the next platform the company wants to dominate could be even bigger: artificial intelligence.

For Google, this doesn’t just mean using AI to improve its own products. (Although it’s certainly doing that). The company wants individuals and small companies around the world to also get on board. It wants to wield influence in the wider AI ecosystem, and to do so has put together an impressive stack of machine learning tools — from software to servers — that mean you can build an AI product from the ground up without ever leaving the Google playpen.

The heart of this offering is Google’s machine learning software TensorFlow. For building AI tools, it’s like the difference between a command line interface and a modern desktop OS; giving users an accessible framework for grappling with their algorithms. It started life as an in-house tool for the company’s engineers to design and train AI algorithms, but in 2015 was made available for anyone to use as open-source software. Since then, it’s been embraced by the AI community (it’s the most popular software of its type on code repository Github), and is used to create custom tools for a whole range of industries, from aerospace to bioengineering.

“There’s hardly a way around TensorFlow these days,” says Samim Winiger, head of machine learning design studio Samim.io. “I use a lot of open source learning libraries, but there’s been a major shift to TensorFlow.”



 

One of Google’s server stacks containing its custom TPU machine learning chips.
Photo: Google

 

Google has made strategic moves to ensure the software is widely used. Earlier this year, for example, it added support for Keras, another popular deep learning framework. According to calculations by the creator of Keras, François Chollet (himself now a Google engineer), TensorFlow was the fastest growing deep learning framework as of September 2016, with Keras in second place. Winiger describes the integration of the two as a “classic tale of Google and how they do it.” He says: “It’s another way that making sure that the entire community converges on their tooling.”

But TensorFlow is also popular for one particularly important reason: it’s good at what it does. “With TensorFlow you get something that scales quickly, works quickly,” James Donkin, a technology manager at UK-based online supermarket Ocado, tells The Verge. He says his team uses a range of machine learning frameworks to create in-house tools for tasks like categorizing customer feedback, but that TensorFlow is often a good place to start. “You get 80 percent of the benefit, and then you might decide to specialize more with other platforms.”

Google offers TensorFlow for free, but it connects easily with the company’s servers for providing data storage or computing power. (“If you use the TensorFlow library it means you can push [products] to Google’s cloud more easily,” says Donkin.) The search giant has even created its own AI-specific chips to power these operations, unveiling the latest iteration of this hardware at this year’s I/O. And, if you want to skip the task of building your own AI algorithms all together, you can buy off-the-shelf components from Google for core tasks like speech transcription and object recognition.

These products and services aren’t necessarily money-makers in themselves, but they other, subtler benefits. They attract talent to Google and help make the company’s in-house software the standard for machine learning. Winiger says these initiatives have helped Google “grab mindshare and make the company’s name synonymous with machine learning.”

Other firms like Amazon, Facebook, and Microsoft also offer their own AI tools, but it’s Google’s that feel pre-eminent. Winiger thinks this is partly down to the company’s capacity to shape the media narrative, but also because of the strong level of support it provides to its users. “There are technical differences between [different AI frameworks], but machine learning communities live off community support and forums, and in that regard Google is winning,” he tells The Verge.

This influence isn’t just abstract, either: it feeds back into Google’s own products. Yesterday, for example, Google announced that Android now has a staggering two billion monthly active users, and to keep the software’s edge, the company is honing it with machine learning. New additions to the OS span the range from tiny tweaks (like smarter text selection) to big new features (like a camera that recognizes what it’s looking at).

But Google didn’t forget to feed the community either, and to complement these announcements unveiled new tools to help developers build AI services that work better on mobile devices. These include a new version of TensorFlow named TensorFlowLite, and an API that will interface with future smartphone chips that have been optimized to work with AI software. Developers can then use these to make better machine learning products for Android devices. Google’s AI empire stretches out a bit further, and Google reaps the benefits.

__

This article and images was originally posted on [The Verge] May 18, 2017 at 06:36AM

by

 

 

 

Google’s AI Invents Sounds Humans Have Never Heard Before

Jesse Engel is playing an instrument that’s somewhere between a clavichord and a Hammond organ—18th-century classical crossed with 20th-century rhythm and blues. Then he drags a marker across his laptop screen. Suddenly, the instrument is somewhere else between a clavichord and a Hammond. Before, it was, say, 15 percent clavichord. Now it’s closer to 75 percent. Then he drags the marker back and forth as quickly as he can, careening though all the sounds between these two very different instruments.“This is not like playing the two at the same time,” says one of Engel’s colleagues, Cinjon Resnick, from across the room. And that’s worth saying. The machine and its software aren’t layering the sounds of a clavichord atop those of a Hammond. They’re producing entirely new sounds using the mathematical characteristics of the notes that emerge from the two. And they can do this with about a thousand different instruments—from violins to balafons—creating countless new sounds from those we already have, thanks to artificial intelligence.
https://soundcloud.com/wired/nsynth-bass-flute

 

Engel and Resnick are part of Google Magenta—a small team of AI researchers inside the internet giant building computer systems that can make their own art—and this is their latest project. It’s called NSynth, and the team will publicly demonstrate the technology later this week at Moogfest, the annual art, music, and technology festival, held this year in Durham, North Carolina.

The idea is that NSynth, which Google first discussed in a blog post last month, will provide musicians with an entirely new range of tools for making music. Critic Marc Weidenbaum points out that the approach isn’t very far removed from what orchestral conductors have done for ages—“the blending of instruments is nothing new,” he says—but he also believes that Google’s technology could push this age-old practice into new places. “Artistically, it could yield some cool stuff, and because it’s Google, people will follow their lead,” he says.

The Boundaries of Sound

Magenta is part of Google Brain, the company’s central AI lab, where a small army of researchers are exploring the limits of neural networks and other forms of machine learning. Neural networks are complex mathematical systems that can learn tasks by analyzing large amounts of data, and in recent years they’ve proven to be an enormously effective way of recognizing objects and faces in photos, identifying commands spoken into smartphones, and translating from one language to another, among other tasks. Now the Magenta team is turning this idea on its head, using neural networks as a way of teaching machines to make new kinds of music and other art.

NSynth begins with a massive database of sounds. Engel and team collected a wide range of notes from about a thousand different instruments and then fed them into a neural network. By analyzing the notes, the neural net—several layers of calculus run across a network of computer chips—learned the audible characteristics of each instrument. Then it created a mathematical “vector” for each one. Using these vectors, a machine can mimic the sound of each instrument—a Hammond organ or a clavichord, say—but it can also combine the sounds of the two.

In addition to the NSynth “slider” that Engel recently demonstrated at Google headquarters, the team has also built a two-dimensional interface that lets you explore the audible space between four different instruments at once. And the team is intent on taking the idea further still, exploring the boundaries of artistic creation. A second neural network, for instance, could learn new ways of mimicking and combining the sounds from all those instruments. AI could work in tandem with AI.

The team has also created a new playground for AI researchers and other computer scientists. They’ve released a research paper describing the NSynth algorithms, and anyone can download and use their database of sounds. For Douglas Eck, who oversees the Magenta team, the hope is that researchers can generate a much wider array of tools for any artist, not just musicians. But not too wide. Art without constraints ceases to be art. The trick will lie in finding the balance between here and the infinite.

__

This article and images was originally posted on [WIRED] May 15, 2017 at 12:57AM

By BUSINESS

 

 

 

Drone Uses AI and 11,500 Crashes to Learn How to Fly

 

“Learning to Fly by Crashing,” a paper from CMU roboticists Dhiraj Gandhi, Lerrel Pinto, and Abhinav Gupta, has such a nice abstract that I’ll just let them explain what this research is all about:

[T]he gap between simulation and real world remains large especially for perception problems. The reason most research avoids using large-scale real data is the fear of crashes! In this paper, we propose to bite the bullet and collect a dataset of crashes itself! We build a drone whose sole purpose is to crash into objects [. . .] We use all this negative flying data in conjunction with positive data sampled from the same trajectories to learn a simple yet powerful policy for UAV navigation.

Cool, let’s get crashing!

One way to think of flying (or driving or walking or any other form of motion) is that success is simply a continual failure to crash. From this perspective, the most effective way of learning how to fly is by getting a lot of experience crashing so that you know exactly what to avoid, and once you can reliably avoid crashing, you by definition know how to fly. Simple, right? We tend not to learn this way, however, because crashing has consequences that are usually quite bad for both robots and people.

The CMU roboticists wanted to see if there are any benefits to using the crash approach instead of the not crash approach, so they sucked it up and let an AR Drone 2.0 loose in 20 different indoor environments, racking up 11,500 collisions over the course of 40 hours of flying time. As the researchers point out, “since the hulls of the drone are cheap and easy to replace, the cost of catastrophic failure is negligible.” Each collision is random, with the drone starting at a random location in the space and then flying slowly forward until it runs into something. After it does, it goes back to its starting point, and chooses a new direction. Assuming it survives, of course.

During this process, the drone’s forward-facing camera is recording images at 30 Hz. Once a collision happens, the images from the trajectory are split into two parts: the part where the drone was doing fine, and the part just before it crashes. These two sets of images are fed into a deep convolutional neural network (with ImageNet-pretrained weights as initialization for the network), which uses them to learn, essentially, whether a given camera image means that going straight is a good idea or not. After 11,500 collisions, the resulting algorithm is able to fly the drone autonomously, even in narrow, cluttered environments, around moving obstacles, and in the midst of featureless white walls and even glass doors. The algorithm that controls the drone is simple: It splits the image from the AR Drone’s forward camera into a left image and a right image, and if one of those two images looks less collision-y than going straight, the drone turns in that direction. Otherwise, it continues moving forward.

How well does this work? It’s usually not as good as a human pilot, except in relatively complex environments, like narrow hallways or hallways with chairs. But compared to a baseline approach using monocular depth estimation, it’s massively better, somewhere between 2x and 10x the performance (in both time in the air and distance flown), depending on the environment. The biggest benefit comes from navigating around featureless walls and glass doors, both of which are notoriously challenging for depth estimation.

The obvious question to ask is whether this method is actually more effective than the alternative, which is teaching a drone to fly through not crashing instead. I’m not sure what the answer is, but the point is that if you allow crashing, the entire learning process can be self-supervised: Just set the drone up in a room and let it do its thing. You’ll have to change the batteries (and the hull, on occasion) but otherwise all of the data collection and learning is completely autonomous. If, on the other hand, you try to teach a drone to fly through not crashing, you have to find a way to make sure that it doesn’t crash. You can do that by learning from a human pilot, or putting it in some environment with a motion capture system and some 3D maps of obstacles and whatnot, but that adds cost and complexity. Crashing is so much easier.

“Learning to Fly by Crashing,” by Dhiraj Gandhi, Lerrel Pinto, and Abhinav Gupta from the Robotics Institute at Carnegie Mellon University, can be read in its entirety at the link below.

[ Paper ]

__

This article and images was originally posted on [IEEE Spectrum Robotics] May 10, 2017 at 09:18AM

By Evan Ackerman

 

 

 

 

 

 

Drug Discovery AI Can Do in a Day What Currently Takes Months

To create a new drug, researchers have to test tens of thousands of compounds to determine how they interact. And that’s the easy part; after a substance is found to be effective against a disease, it has to perform well in three different phases of clinical trials and be approved by regulatory bodies.

It’s estimated that, on average, one new drug coming to market can take 1,000 people, 12-15 years, and up to $1.6 billion.

There has to be a better way—and now it seems there is.

Last week, researchers published a paper detailing an artificial intelligence system made to help discover new drugs, and significantly shorten the amount of time and money it takes to do so.

The system is called AtomNet, and it comes from San Francisco-based startup AtomWise. The technology aims to streamline the initial phase of drug discovery, which involves analyzing how different molecules interact with one another—specifically, scientists need to determine which molecules will bind together and how strongly. They use trial and error and process of elimination to analyze tens of thousands of compounds, both natural and synthetic.

AtomNet takes the legwork out of this process, using deep learning to predict how molecules will behave and how likely they are to bind together. The software teaches itself about molecular interaction by identifying patterns, similar to how AI learns to recognize images.

Remember the 3D models of atoms you made in high school, where you used pipe cleaners and foam balls to represent the connections between protons, neutrons and electrons? AtomNet uses similar digital 3D models of molecules, incorporating data about their structure to predict their bioactivity.

As AtomWise COO Alexander Levy put it, “You can take an interaction between a drug and huge biological system and you can decompose that to smaller and smaller interactive groups. If you study enough historical examples of molecules…you can then make predictions that are extremely accurate yet also extremely fast.”

“Fast” may even be an understatement; AtomNet can reportedly screen one million compounds in a day, a volume that would take months via traditional methods.

AtomNet can’t actually invent a new drug, or even say for sure whether a combination of two molecules will yield an effective drug. What it can do is predict how likely a compound is to work against a certain illness. Researchers then use those predictions to narrow thousands of options down to dozens (or less), focusing their testing where there’s more likely to be positive results.

The software has already proven itself by helping create new drugs for two diseases, Ebola and multiple sclerosis. The MS drug has been licensed to a British pharmaceutical company, and the Ebola drug is being submitted to a peer-reviewed journal for additional analysis.

While AtomNet is a promising technology that will make discovering new drugs faster and easier, it’s worth noting that the future of medicine is also moving towards a proactive rather than reactive approach; rather than solely inventing drugs to cure sick people, focus will shift to carefully monitoring our health and taking necessary steps to keep us from getting sick in the first place.

Last year, the Chan Zuckerberg Initiative donated $3 billion in a pledge to “cure all diseases.” It’s an ambitious and somewhat quixotic goal, but admirable nonetheless. In another example of the movement towards proactive healthcare, the XPRIZE foundation recently awarded $2.5 million for a device meant to facilitate home-based diagnostics and personal health monitoring. Proactive healthcare technology is likely to keep advancing and growing in popularity.

That doesn’t mean reactive healthcare shouldn’t advance alongside it; fifty or one hundred years from now, people will still be getting sick and will still need medicine to help cure them. AtomNet is the first software of its kind, and it may soon see others following in its footsteps in the effort to apply AI to large-scale challenges.

Image Credit: Shutterstock

__

This article and images was originally posted on [Singularity Hub] May 7, 2017 at 04:01AM

Amazon to control 70 percent of the voice-controlled speaker market this year

Amazon is dominating the voice-controlled speaker market, according to a new forecast from eMarketer out this morning. The maker of the Echo-branded speakers will have 70.6 percent of all voice-enabled speaker users in the U.S. this year – well ahead of Google Home’s 23.8 percent and other, smaller players like Lenovo, LG, Harmon Kardon, and Mattel, who combined only account for 5.6 percent of users.

The new report backs up another from VoiceLabs released in January, which also found that Amazon was leading the voice-first device market, thanks to Echo’s popularity.

While the market itself is not expected to be a winner-take-all scenario, competitors like Amazon and Google will win entire homes, as most consumers have said they wouldn’t consider buying a competing device once they already own one voice-controlled speaker.

Emarketer says Amazon’s share will fall somewhat as Google gets its Home devices into more consumers’ hands, but Amazon will lead the market for the “foreseeable future.”

This year, 35.6 million Americans will use a voice-activated device at least once per month, the firm estimates – a 128.9 percent over 2016. (The company is not counting voice assistants on smartphones in this part of its forecast – only those digital assistants on standalone voice-enabled speakers.)

However, the wider voice assistant market – which includes Siri, Alexa, Google Now and Microsoft’s Cortana operating on any device –  is expected to grow 23.1 percent this year, eMarketer says. In 2017, 60.5 million U.S. users will use one of these assistants at least once per month. That’s over a quarter (27.5%) of smartphone users, or nearly one in five Americans, the report notes.

Usage is heaviest among a younger demographic – those between 25 and 34, make up 26.3 percent of virtual assistant users. In addition, more than one-third of millennials (33.5%) will use virtual assistants in 2017.

“Older millennials are the core users of virtual assistants, mainly due to their demand for functionality over entertainment,”  eMarketer’s VP of Forecasting, Martín Utreras, says.

Amazon doesn’t disclose Echo sales figures, but said in February that Echo family sales are up over 9x compared to last season. VoiceLabs estimated in January there are over 7 million devices in customers’ homes; Consumer Intelligence Research Partners (CIRP) around the same time put it higher at 8.2 million; and Morgan Stanley believes it was over 11 million before the 2017 holidays.

Amazon’s strategy with Alexa is to allow the assistant to extend beyond just voice-controlled speakers it manufacturers. The company has also included the assistant in its Amazon mobile shopping app, and has made it available to third-parties for use in their own hardware and software applications.

With its dominant market share, Amazon’s Alexa could bring in $10 billion in revenues by 2020, RBC Capital Markets had also estimated in March. This figure included Alexa device sales, voice shopping sales, and platform revenue. The firm forecasted there will be 60 million Alexa devices sold in 2020, bringing the install base to 128 million.

__

This article and images was originally posted on [TechCrunch] May 8, 2017 at 05:20AM

by

 

 

 

Nvidia Lets You Peer Inside the Black Box of Its Self-Driving AI

Nvidia has developed a self-driving AI that shows you how it works.

As we explained in our latest cover story, “The Dark Secret at the Heart of AI,” some of the most powerful machine learning techniques available result in software that is almost completely opaque, even to the engineers that build it. Approaches that provide some clues as to how an AI works will, therefore, be hugely important for building trust in a technology that looks set to revolutionize everything from medicine to manufacturing.

Nvidia provides chips that are ideal for deep learning, an especially powerful machine learning technique (see “10 Breakthrough Technologies 2013: Deep Learning”).

Nvidia’s neural network software highlights the areas it’s focusing on as it makes driving decisions.

The chipmaker has also been developing systems that demonstrate how an automaker might apply deep learning to autonomous driving. This includes a car that is controlled entirely by a deep learning algorithm. Amazingly, the vehicle’s computer isn’t given any rules to follow—it simply matches input from several video cameras to the behavior of a human driver, and figures out for itself how it should drive. The only catch is that the system is so complex that it’s difficult to untangle how it actually works.

But Nvidia is working to open this black box. It has developed a way to visually highlight what the system is paying attention to. As explained in a recently published paper, the neural network architecture developed by Nvidia’s researchers is designed so that it can highlight the areas of a video picture that contribute most strongly to the behavior of the car’s deep neural network. Remarkably, the results show that the network is focusing on the edges of roads, lane markings, and parked cars—just the sort of things that a good human driver would want to pay attention to.

“What’s revolutionary about this is that we never directly told the network to care about these things,” Urs Muller, Nvidia’s chief architect for self-driving cars, wrote in a blog post.

It isn’t a complete explanation of how the neural network reasons, but it’s a good start. As Muller says: “I can’t explain everything I need the car to do, but I can show it, and now it can show me what it learned.”

This sort of approach could become increasingly important as deep learning is applied to just about any problem involving large quantities of data, including critical areas like medicine, finance, and military intelligence.

A handful of academic researchers are exploring the issue as well. For example, Jeff Clune at the University of Wyoming and Carlos Guestrin at the University of Washington (and Apple) have found ways of highlighting the parts of images that classification systems are picking up on. And Tommi Jaakola and Regina Barzilay at MIT are developing ways to provide snippets of text that help explain a conclusion drawn from large quantities of written data.

The Defense Advanced Projects Research Agency (DARPA), which does long-term research for the US military, is funding several similar research efforts through a program it calls Explainable Artificial Intelligence (XAI).

Beyond the technical specifics, though, it’s fascinating to consider how this compares to human intelligence. We do all sorts of things we can’t explain fully, and the explanations we concoct are often only approximations, or “stories” about what’s going on. Given the opacity of today’s increasingly complex machine learning methods, we may someday be forced to accept such explanations from AI, too.

(Sources: Nvidia, “The Dark Secret at the Heart of AI”, “The U.S. Military Wants Its Autonomous Machines to Explain Themselves”)

__

This article and images was originally posted on [New on MIT Technology Review] May 3, 2017 at 05:46AM

by Will Knight

 

 

Bixby voice arrives on Korean Galaxy S8 today

Galaxy S8 and S8+ owners in Samsung’s home market are the first to get a fully baked Bixby experience.

Bixby, Samsung’s new AI service on the Galaxy S8, was missing its most important feature at launch. Bixby’s voice commands — a central reason for the service having its own dedicated hardware button — wasn’t operational out of the box. However from today, Korean Galaxy S8 owners can get acquainted with Samsung’s AI-based trickery. ZDNet reports that Samsung flipped the switch on Bixby voice at 1pm KST on Monday (11pm EST Sunday).

Finally, a reason to push the Bixby button.

Initially, Bixby’s voice commands only work in a handful of Samsung’s own apps — Gallery, Settings, Camera and Reminders. Expect more apps and services to be added in the coming months, as Bixby voice eventually hits more countries around the world. Next up is the U.S. later this spring, with other English-speaking locales likely to follow.

As for languages besides English and Korean, those territories may be in for a much longer wait. Samsung has said that it expects to roll out Bixby voice in Germany in Q4, by which time the Galaxy Note 8 launch should be near. Other European languages might expect a similar release timeframe.

With so few apps supported at launch, it’s not like Galaxy S8 owners in other countries are missing out on much. As we’ve said before, Bixby is going to be a slow burn, and it’ll take several months for the voice features to expand into new apps anyway.

__

This article and images was originally posted on [Android Central – Android Forums, News, Reviews, Help and Android Wallpapers] April 30, 2017 at 09:07PM

BY ALEX DOBIE

 

 

 

Grand Theft Auto 5 is being used to teach driverless cars

Grand Theft Auto 5 is being used as a simulation to test the AI of driverless cars.

As reported by Bloomberg, Rockstar’s silly sandbox game happens to provide the best simulation we have at the moment. After all, it consists of 262 vehicles, 14 weather conditions, over a thousand unpredictable NPCs roaming about, along with plenty of bridges, tunnels and traffic signals.

According to Princeton University professor of operations research and financial engineering Alain Kornhauser, GTA5 is “the richest virtual environment that we could extract data from.”

While using a video game to train the AI of semi-sentient vehicles may seem outlandish, it’s safer than testing them in the real world, plus it offers engineers greater control over testing each variable. A simulation is a more controlled environment, after all, which is important for running experiments.

“Just relying on data from the roads is not practical,” said head of self-driving car startup Nio, Davide Bacchet. “With simulation, you can run the same scenario over and over again for infinite times, then test it again.”

Indeed, it’s best for engineers to test self-driving cars under the most hazardous of conditions, and what could be worse than this:

 

But hey, if a driverless car knows how to safely navigate this mess, that bodes well for it.

__

Join our fans by liking us on Facebook, or follow us on Twitter, Google+, feedly, flipboardand Instagram.

This article and images was originally posted on Eurogamer

By Jeffrey Matulef

 

 

 

AI can predict heart attacks more accurately than doctors

An estimated 20 million people die each year due to cardiovascular disease. Luckily, a team of researchers from the University of Nottingham in the UK have developed a machine-learning algorithm that can predict your likelihood of having a heart attack or stroke as well as any doctor.

The American College of Cardiology/American Heart Association (ACC/AHA) has developed a series of guidelines for estimating a patient’s cardiovascular risk which is based on eight factors including age, cholesterol level and blood pressure. On average, this system correctly guesses a person’s risk at a rate of 72.8 percent.

That’s pretty accurate but Stephen Weng and his team set about to make it better. They built four computer learning algorithms, then fed them data from 378,256 patients in the United Kingdom. The systems first used around 295,000 records to generate their internal predictive models. Then they used the remaining records to test and refine them. The algorithms results significantly outperformed the AAA/AHA guidelines, ranging from 74.5 to 76.4 percent accuracy. The neural network algorithm tested highest, beating the existing guidelines by 7.6 percent while raising 1.6 percent fewer false alarms.

Out of the 83,000 patient set of test records, this system could have saved 355 extra lives. Interestingly, the AI systems identified a number of risk factors and predictors not covered in the existing guidelines, like severe mental illness and the consumption of oral corticosteroids. “There’s a lot of interaction in biological systems,” Weng told Science. “That’s the reality of the human body. What computer science allows us to do is to explore those associations.”

Source: Science

__

Join our fans by liking us on Facebook, or follow us on Twitter, Google+, feedly, flipboardand Instagram.

This article and images was originally posted on Engadget

by Andrew Tarantola, @terrortola

 

 

 

 

 

OpenAI Just Beat Google DeepMind at Atari With an Algorithm From the 80s

AI research has a long history of repurposing old ideas that have gone out of style. Now researchers at Elon Musk’s open source AI project have revisited “neuroevolution,” a field that has been around since the 1980s, and achieved state-of-the-art results.

The group, led by OpenAI’s research director Ilya Sutskever, has been exploring the use of a subset of algorithms from this field, called “evolution strategies,” which are aimed at solving optimization problems.

Despite the name, the approach is only loosely linked to biological evolution, the researchers say in a blog post announcing their results. On an abstract level, it relies on allowing successful individuals to pass on their characteristics to future generations. The researchers have taken these algorithms and reworked them to work better with deep neural networks and run on large-scale distributed computing systems.

“To validate their effectiveness, they then set them to work on a series of challenges seen as benchmarks for reinforcement learning.”

To validate their effectiveness, they then set them to work on a series of challenges seen as benchmarks for reinforcement learning, the technique behind many of Google DeepMind’s most impressive feats, including beating a champion Go player last year.

One of these challenges is to train the algorithm to play a variety of computer games developed by Atari. DeepMind made the news in 2013 when it showed it could use Deep Q-Learning—a combination of reinforcement learning and convolutional neural networks—to successfully tackle seven such games. The other is to get an algorithm to learn how to control a virtual humanoid walker in a physics engine.

To do this, the algorithm starts with a random policy—the set of rules that govern how the system should behave to get a high score in an Atari game, for example. It then creates several hundred copies of the policy—with some random variation—and these are tested on the game.

These policies are then mixed back together again, but with greater weight given to the policies that got the highest score in the game. The process repeats until the system comes up with a policy that can play the game well.

“In one hour training on the Atari challenge, the algorithm reached a level of mastery that took a [DeepMind] reinforcement-learning system…a whole day to learn.”

In one hour training on the Atari challenge, the algorithm reached a level of mastery that took a reinforcement-learning system published by DeepMind last year a whole day to learn. On the walking problem the system took 10 minutes, compared to 10 hours for Google’s approach.

One of the keys to this dramatic performance was the fact that the approach is highly “parallelizable.” To solve the walking simulation, they spread computations over 1,440 CPU cores, while in the Atari challenge they used 720.

This is possible because it requires limited communication between the various “worker” algorithms testing the candidate policies. Scaling reinforcement algorithms like the one from DeepMind in the same way is challenging because there needs to be much more communication, the researchers say.

The approach also doesn’t require backpropagation, a common technique in neural network-based approaches, including deep reinforcement learning. This effectively compares the network’s output with the desired output and then feeds the resulting information back into the network to help optimize it.

The researchers say this makes the code shorter and the algorithm between two and three times faster in practice. They also suggest it will be particularly suited to longer challenges and situations where actions have long-lasting effects that may not become apparent until many steps down the line.

The approach does have its limitations, though. These kinds of algorithms are usually compared based on their data efficiency—the number of iterations required to achieve a specific score in a game, for example. On this metric, the OpenAI approach does worse than reinforcement learning approaches, although this is offset by the fact that it is highly parallelizable and so can carry out iterations more quickly.

For supervised learning problems like image classification and speech recognition, which currently have the most real-world applications, the approach can also be as much as 1,000 times slower than other approaches that use backpropagation.

Nevertheless, the work demonstrates promising new applications for out-of-style evolutionary approaches, and OpenAI is not the only group investigating them. Google has been experimenting on using similar strategies to devise better image recognition algorithms. Whether this represents the next evolution in AI we will have to wait and see.

Image Credit: Shutterstock

__

Join our fans by liking us on Facebook, or follow us on Twitter, Google+, feedly, flipboardand Instagram.

This article and images was originally posted on Singularity Hub