Machine learning vs AI
These two terms are often used interchangeably, but they are fundamentally different technologies
Technology is evolving at a rapid pace. It’s been made to, to in a way, with the global pandemic forcing businesses to digitally transform faster than they ever imagined they would within a short span of time.
The rate of acceleration, when it comes to this level of innovation, knows no bounds. The constant development of features and functions means it can be hard to keep up, and you might ask yourself whether the new tech is making your life easier, or adding complications.
Even understanding what the new technology is, and what it does, can be confusing. Never more so than when two of the top new trending technologies, artificial intelligence (AI) and machine learning (ML) appear, to those of us less familiar with them, to do the same thing.
It’s true that they’re mentioned together more often than not, particularly in commercial settings or in films, and essentially one can’t happen without the other, but by definition, there are clear differences.
Most people, especially if they’ve seen the 2001 Steven Spielberg movie, think AI is the combination of humans and machines, however, it is a much broader term for anything that enables computers to act like humans.
Popular known examples of this include smart speakers, like Apple’s Siri or Amazon's Alexa, or even virtual assistants or chatbots on your favourite retail sites. However, other more wide-ranging uses include more business-related applications, like statistical analysis for pricing models, and even fraud protection.
Machine learning, on the other hand, is a type of AI. Essentially it enables machines to learn from data, but it has limited scope, unlike AIML itself also incorporates various subdivisions such as reinforcement learning and deep learning.
What's the difference between ML and AI?
The history of AI is a long one. For thousands of years, humans have dreamt of machines that could 'come to life', behaving and thinking as humans do. There was a time when early computers, due to their 'logical' nature, were also considered a type of artificial intelligence.
In its current manifestation, however, the idea of AI can trace its history to British computer scientist and World War II codebreaker Alan Turing. He proposed a test, which he called the imitation game but is more commonly now known as the Turing Test, where one individual converses with two others, one of which is a machine, through a text-only channel. If the interrogator is unable to tell the difference between the machine and the person, the machine is considered to have "passed" the test.
This basic concept is referred to as 'general AI' and is generally considered to be something that researchers have yet to fully achieve.
However, 'narrow' or 'applied' AI has been far more successful at creating working models. Rather than attempt to create a machine that can do everything, this field attempts to create a system that can perform a single task as well as, if not better than, a human.
It's within this narrow AI discipline that the idea of machine learning first emerged, as early as the middle of the twentieth century. First defined by AI pioneer Arthur Samuel in a 1959 academic paper, ML represents "the ability to learn without being explicitly programmed".
Uses and applications
Interest in ML has waxed and waned over the years, but with data becoming an increasingly important part of business strategy, it's fallen back into favour as organisations seek ways to analyse and make use of the vast quantities of information they collect on an almost constant basis.
When this data is put into a machine learning program, the software not only analyses it but learns something new with each new dataset, becoming a growing source of intelligence. This means the insights that can be learnt from data sources become more advanced and more informative, helping companies develop their business in line with customer expectations.
One application of ML is in a recommendation engine, like Facebook's newsfeed algorithm, or Amazon's product recommendation feature. ML can analyse how many people are liking, commenting on or sharing posts or what other people are buying that have similar interests. It will then show the post to others the system thinks will like it.
ML is also particularly useful for image recognition, using humans to identify what's in a picture as a kind of programming and then using this to autonomously identify what's in a picture. For example, machine learning can identify the distribution of the pixels used in a picture, working out what the subject is.
Enterprises are now turning to ML to drive predictive analytics, as big data analysis becomes increasingly widespread. The association with statistics, data mining and predictive analysis have become dominant enough for some to argue that machine learning is a separate field from AI.
The reason for this is that AI technology, such as natural language processing or automated reasoning, can be done without having the capability for machine learning. It is not always necessary for ML systems to have other features of AI.
There are hundreds of use cases for AI, and more are becoming apparent as companies adopt artificial intelligence to tackle business challenges.
One of the most common uses of AI is for automation in cyber security. For example, AI algorithms can be programmed to detect threats that may be difficult for a human to spot, such as subtle changes in user behaviour or an unexplained increase in the amount of data being transferred to and from a particular node (such as a computer or sensor). In the home, assistants like Google Home or Alexa can help automate lighting, heating and interactions with businesses through chatbots.
There are well-founded fears that AI will replace human job roles, such as data input, at a faster rate than the job market will be able to adapt to. Author and venture capitalist Kai-Fu Lee, who has worked at both Apple and Google and earned a PhD from Carnegie Mellon for the development of an advanced speech recognition AI, warned in 2019 that "many jobs that seem a little bit complex, a chef, a waiter, a lot of things, will become automated."
"We will have automated stores, automated restaurants and all together, in 15 years, that's going to displace about 40% of jobs in the world."
Confusing AI and ML
To make matters more confusing when it comes to naming and identifying these terms, there are a number of other terms thrown into the hat. These include artificial neural networks, for instance, which process information in a way that mimics neurons and synapses in the human mind. This technology can be used for machine learning; although not all neural networks are AI or ML, and not all ML programmes use underlying neural networks.
As this is a developing field, terms are popping in and out of existence all the time and the barriers between the different areas of AI are still quite permeable. As the technology becomes more widespread and more mature, these definitions will likely also become more concrete and well known. On the other hand, if we develop generalised AI, all these definitions may suddenly cease to be relevant.
Next-generation time series: Forecasting for the real world, not the ideal world
Solve time series problems with AIFree download
The future of productivity
Driving your business forward with Microsoft Office 365Free download
How to plan for endpoint security against ever-evolving cyber threats
Safeguard your devices, data, and reputationFree download
A quantitative comparison of UPS monitoring and servicing approaches across edge environments
Effective UPS fleet managementFree download