What is machine learning?

ML is no longer the work of science fiction as machines are getting cleverer by the day

AI artificial intelligence

Machine learning (ML) is the process of teaching a computer system to make predictions based on a set of data. By feeding a system a series of trial and error scenarios, machine learning researchers strive to create artificially intelligent systems that can analyse data, answer questions, and make decisions on their own. 

Machine learning often uses algorithms based on test data, which assist with inference and pattern recognition in future decisions, removing the need for explicit instructions from humans that traditional computer software requires. 

What is machine learning?

Machine learning relies on a large amount of data, which is fed into algorithms in order to produce a model off of which the system predicts its future decisions. For example, if the data you’re feeding into the system is which fruits you’ve eaten for lunch every day for a year, you would be able to use a prediction algorithm to analyse the different fruits and build a prediction model for which fruits you were likely to eat when in the following year. 

The process is based on trial and error scenarios, usually using more than one algorithm. These algorithms are classed as linear models, non-linear models, or even neural networks. They will be ultimately dependent on the set of data you’re working with and the question you’re trying to answer.

How do machine learning algorithms work?

Machine learning algorithms learn and improve over time using data, and do not require human instruction. The algorithms are split into three types: supervised, unsupervised, and reinforcement learning. Each type of learning has a different purpose and enables data to be used in different ways.

Supervised learning

Supervised learning involves labelled training data, which is used by an algorithm to learn the mapping function that turns input variables into an output variable to solve equations. Within this are two types of supervised learning: classification, which is used to predict the outcome of a given sample when the output is in the form of a category, and regression, which is used to predict the outcome of a given sample when the output variable is a real value, such as a 'salary' or a 'weight'.

An example of a supervised learning model is the K-Nearest Neighbors (KNN) algorithm, which is a method of pattern recognition. KNN essentially involves using a chart to reach an educated guess on the classification of an object based on the spread of similar objects nearby.

In the chart above, the green circle represents an as-yet unclassified object, which can only belong to one of two possible categories: blue squares or red triangles. In order to identify what category it belongs to, the algorithm will analyse what objects are nearest to it on the chart in this case, the algorithm will reasonably assume that the green circle should belong to the red triangle category. 

Unsupervised learning

Unsupervised learning models are used when there is only input variables and no corresponding output variables. It uses unlabelled training data to model the underlying structure of the data.

There are three types of unsupervised learning algorithms: association, which is extensively used in market-basket analysis; clustering, which is used to match samples similar to objects within another cluster; and dimensionality reduction, which is used to trim the number of variables within a data set while keeping its important information intact.

Reinforcement learning

Reinforcement learning allows an agent to decide its next action based on its current state by learning behaviours that will maximize a reward. It's often used in gaming environments where an algorithm is provided with the rules and tasked with solving the challenge in the most efficient way possible. The model will start out randomly at first, but over time, through trial and error, it will learn where and when it needs to move in the game to maximize points.

In this type of training, the reward is simply a state associated with a positive outcome. For example, an algorithm will be 'rewarded' with a task completion if it is able to keep a car on a road without hitting obstacles. 

Why is machine learning useful?

In essence, ML solves the problem of too much data; we have so much information being generated by people, actions, events, computers and gadgets that to learn anything from them is virtually impossible for humans. Within medical analysis, finding patterns in thousands of MRI scans would take a human many hours, days or weeks to complete, but a machine can ingest that information and spot the patterns in seconds if they are correctly labelled.

Where is machine learning used?

One of the simplest and most successful examples of machine learning is something we use every day - Google Search. The search engine is powered by many ML algorithms that read and analyse the text you put in, tailoring the results based on your search history and online habits. For instance, if you type in 'Java' you'll either get results around the programming language or for coffee surfaced more frequently, depending on which it has determined you'll prefer.

Many of our future technological advancements depend on the development of machine learning, such as driverless cars and smart cities. Many of the systems to power smart cities are entering the public space, such as facial recognition systems where ML algorithms are taught to recognise patterns in images and identify objects based on their characteristics. This, however, has proven to be a controversial use of ML, particularly as it isn't always accurate and often involves some sort of regular surveillance of citizens.

Data bias

As machine learning improves and is used in more technologies, the worry about embedding bias into critical and public-facing software grows. ML applications are dependent on data and it's this data that can be the source of bias. For example, if a company that wants to hire more diversely, but uses its current employee's CVs, by default its machine learning program will only look for more of the same.

Related Resource

Intelligent process automation

Boosting bots with AI and machine learning

Download now

It's this type of application of machine learning that has governments worried and, as such, many are resorting to enforcing rules and regulations to combat this issue. The UK's Centre for Data Ethics and Innovation (CDEI) announced it was to join forces with the Cabinet Office's Race Disparity Unit to investigate potential bias in algorithmic decision-making. Likewise, the US government is to pilot diversity regulations for work on AI that reduces the risk of sexual and racial bias within computer systems.

Image by Antti Ajanki AnAj / CC BY-SA 3.0

Most Popular

The benefits of workload optimisation

The benefits of workload optimisation

16 Jul 2021
Samsung Galaxy S21 5G review: A rose-tinted experience
Mobile Phones

Samsung Galaxy S21 5G review: A rose-tinted experience

14 Jul 2021
Six ways boards can step up support for cyber security
Business strategy

Six ways boards can step up support for cyber security

22 Jul 2021