Upgrade & Secure Your Future with DevOps, SRE, DevSecOps, MLOps!

We spend hours on Instagram and YouTube and waste money on coffee and fast food, but won’t spend 30 minutes a day learning skills to boost our careers.
Master in DevOps, SRE, DevSecOps & MLOps!

Learn from Guru Rajesh Kumar and double your salary in just one year.

Get Started Now!

Google AI Expert: Machine Learning Is No Better Than Alchemy

Source – livescience.com

A prominent researcher of machine learning and artificial intelligence is arguing that his field has strayed out of the bounds of science and engineering and into “alchemy.” And he’s offering a route back.

Ali Rahimi, who works on AI for Google, said he thinks his field has made amazing progress, but suggested there’s something rotten in the way it’s developed. In machine learning, a computer “learns” via a process of trial and error. The problem in a talk presented at an A.I. conference is that researchers who work in the field — when a computer “learns” due to a process of trial and error — not only don’t understand exactly how their algorithms learn, but they don’t understand how the techniques they’re using to build those algorithms work either, Rahimi suggested in a talk presented at an AI conference covered recently by Matthew Hutson for Science magazine.

Back in 2017, Rahimi sounded the alarm on the mystical side of artificial intelligence: “We produce stunningly impressive results,” he wrote in a blog. “Self-driving cars seem to be around the corner; artificial intelligence tags faces in photos, transcribes voicemails, translates documents and feeds us ads. Billion-dollar companies are built on machine learning. In many ways, we’re in a better spot than we were 10 years ago. In some ways, we’re in a worse spot.”

Rahimi, as Hutson reported, showed that many machine-learning algorithms contain tacked-on features that are essentially useless, and that many algorithms work better when those features are stripped away. Other algorithms are fundamentally broken and work only because of a thick crust of ad-hoc fixes piled on top of the original program.

This is, at least in part, the result of a field that’s gotten used to a kind of random, trial-and-error methodology, Rahimi argued in that blog. Under this process, researchers don’t understand at all why one attempt at solving a problem worked and another failed. People implement and share techniques that they don’t remotely understand.

Folks who follow AI might be reminded of the “black box” problem, Hutson noted in his article — the tendency of AI programs to solve problems in ways that their human creators don’t understand. But the current issue is different: Researchers not only don’t understand their AI programs’ problem-solving techniques, Rahimi said, but they don’t understand the techniques they used to build those programs in the first place either. In other words, the field is more like alchemy than a modern system of research, he said.

“There’s a place for alchemy. Alchemy worked,” Rahimi wrote.

“Alchemists invented metallurgy, ways to make medication, dy[e]ing techniques for textiles, and our modern glass-making processes. Then again, alchemists also believed they could transmute base metals into gold and that leeches were a fine way to cure diseases.”

In his more recent talk (and accompanying paper) at the International Conference on Learning Representations in Vancouver, Canada, Rahimi and several colleagues proposed a number of methods and protocols that could move machine learning beyond the world of alchemy. Among them: evaluating new algorithms in terms of their constituent parts, deleting parts of them one at a time and testing if the overall programs still work, and performing basic “sanity tests” on the results that the algorithms produce.

That’s all because AI, Rahimi argued in his 2017 blog, has become too important in society to be developed in such a slapdash fashion.

“If you’re building photo-sharing services, alchemy is fine,” he wrote. “But we’re now building systems that govern health care and our participation in civil debate. I would like to live in a world whose systems are built on rigorous, reliable, verifiable knowledge and not on alchemy.”

Related Posts

What is AIOps?

AIOps, short for Artificial Intelligence for IT Operations, is a practice that combines artificial intelligence (AI) and machine learning (ML) technologies with traditional IT operations to enhance Read More

Read More

What is Natural Language Processing (NLP) tools?

Introduction to Natural Language Processing (NLP) Tools If you’ve ever asked Siri a question or talked to Alexa, you’ve used Natural Language Processing (NLP) tools. In essence, Read More

Read More

What are Emotion Detection Tools and Why Emotion Detection Tools are Important?

What are Emotion Detection Tools? Emotion detection tools are a type of technology that analyses human facial expressions, voice tone, and body language to determine the emotional Read More

Read More

What is Sentiment Analysis and what are the Types of Sentiment Analysis and its Important?

Introduction to Sentiment Analysis If you’re a business owner, marketer, or just someone who’s curious about what people think about your brand, then you’ve probably heard of Read More

Read More

What is Object Detection and Why is Object Detection Important?

Introduction to Object Detection Tools Object detection is the process of identifying and locating objects of interest in an image or video. Object detection tools are software Read More

Read More

What is Face Recognition and Why is Face Recognition Important?

Introduction to Face Recognition Tools We’ve all heard of facial recognition technology, but what exactly is it and why is it important? From unlocking your phone with Read More

Read More
Subscribe
Notify of
guest
3 Comments
Oldest
Newest Most Voted
Inline Feedbacks
View all comments
3
0
Would love your thoughts, please comment.x
()
x