The Complete Beginners' Guide to Artificial Intelligence

Ten years ago, if you mentioned the term “artificial intelligence” in a boardroom there’s a good chance you would have been laughed at. For most people it would bring to mind sentient, sci-fi machines such as 2001: A Space Odyssey’s HAL or Star Trek’s Data.
Today it is one of the hottest buzzwords in business and industry. AI technology is a crucial lynchpin of much of the digital transformation taking place today as organizations position themselves to capitalize on the ever-growing amount of data being generated and collected.

So how has this change come about? Well partly it is due to the Big Data revolution itself. The glut of data has led to intensified research into ways it can be processed, analyzed and acted upon. Machines being far better suited than humans to this work, the focus was on training machines to do this in as “smart” a way as is possible.

This increased interest in research in the field – in academia, industry and among the open source community which sits in the middle – has led to breakthroughs and advances that are showing their potential to generate tremendous change. From healthcare to self-driving cars to predicting the outcome of legal cases, no one is laughing now!

What is Artificial Intelligence

The concept of what defines AI has changed over time, but at the core there has always been the idea of building machines which are capable of thinking like humans.

After all, human beings have proven uniquely capable of interpreting the world around us and using the information we pick up to effect change. If we want to build machines to help us do this more efficiently, then it makes sense to use ourselves as a blueprint.

AI, then, can be thought of as simulating the capacity for abstract, creative, deductive thought – and particularly the ability to learn which this gives rise to – using the digital, binary logic of computers.

Research and development work in AI is split between two branches. One is labelled “applied AI” which uses these principles of simulating human thought to carry out one specific task. The other is known as “generalized AI” – which seeks to develop machine intelligence that can turn their hands to any task, much like a person.

Research into applied, specialized AI is already providing breakthroughs in fields of study from quantum physics where it is used to model and predict the behavior of systems comprised of billions of subatomic particles, to medicine where it being used to diagnose patients based on genomic data.

What Is Machine Learning - A Complete Beginner's Guide In 2017

Difference Between Artificial Intelligence And Machine Learning


In industry, it is employed in the financial world for uses ranging from fraud detection to improving customer service by predicting what services customers will need. In manufacturing it is used to manage workforces and production processes as well as for predicting faults before they occur, therefore enabling predictive maintenance.

In the consumer world more and more of the technology we are adopting into our everyday lives is becoming powered by AI – from smartphone assistants like Apple’s Siri and Google’s Google Assistant, to self-driving and autonomous cars which many are predicting will outnumber manually driven cars within our lifetimes.

Generalized AI is a bit further off – to carry out a complete simulation of the human brain would require both a more complete understanding of the organ than we currently have, and more computing power than is commonly available to researchers. But that may not be the case for long, given the speed with which computer technology is evolving. A new generation of computer chip technology is known as neuromorphic processors are being designed to more efficiently run brain-simulator code. And systems such as IBM’s Watson cognitive computing platform use high-level simulations of human neurological processes to carry out an ever-growing range of tasks without being specifically taught how to do them.

What are the key developments in AI?

All of these advances have been made possible due to the focus on imitating human thought processes. The field of research which has been most fruitful in recent years is what has become known as “machine learning”. In fact, it’s become so integral to contemporary AI that the terms “artificial intelligence” and “machine learning” are sometimes used interchangeably.
However, this is an imprecise use of language, and the best way to think of it is that machine learning represents the current state-of-the-art in the wider field of AI. The foundation of machine learning is that rather than have to be taught to do everything step by step, machines, if they can be programmed to think like us, can learn to work by observing, classifying and learning from its mistakes, just like we do.

The application of neuroscience to IT system architecture has led to the development of artificial neural networks – and although work in this field has evolved over the last half century it is only recently that computers with adequate power have been available to make the task a day-to-day reality for anyone except those with access to the most expensive, specialized tools.
Perhaps the single biggest enabling factor has been the explosion of data which has been unleashed since mainstream society merged itself with the digital world. This availability of data – from things we share on social media to machine data generated by connected industrial machinery – means computers now have a universe of information available to them, to help them learn more efficiently and make better decisions.

What is the future of AI?

That depends on who you ask, and the answer will vary wildly!
Real fears that development of intelligence which equals or exceeds our own, but has the capacity to work at far higher speeds, could have negative implications for the future of humanity have been voiced, and not just by apocalyptic sci-fi such as The Matrix or The Terminator, but respected scientists like Stephen Hawking.

Even if robots don’t eradicate us or turn us into living batteries, a less dramatic but still nightmarish scenario is that automation of labour (mental as well as physical) will lead to profound societal change – perhaps for the better, or perhaps for the worse.

This understandable concern has led to the foundation last year, by a number of tech giants including Google, IBM, Microsoft, Facebook and Amazon, of the Partnership in AI. This group will research and advocate for ethical implementations of AI, and to set guidelines for future research and deployment of robots and AI.

Via Fobes contributed by Bernard Marr  

Related Posts:

What Is Machine Learning - A Complete Beginner's Guide In 2017

Difference Between Artificial Intelligence And Machine Learning

Big Data Terminology: 16 Key Concepts Everyone Should Understand

A Complete Beginner's Guide To Blockchain


No comments