*originally published at CSB Tech Blog
Over the past decade, AI inspired innovations have begun to mature, becoming the hottest tech trend of the new millennium. During that time, computer vision improved by leaps and bounds, along with machine translation, speech recognition, and many types of data analytics. While we’ve experienced many dramatic improvements in our online experience, machine learning has been working quietly in the background. These advances are made even more dramatic in that they seem to have occurred overnight. Although the applications of AI research do not possess an independent intelligence, with today’s processing capabilities their capacity for automation is unparalleled.
Deep
learning is a branch of machine learning (ML) that came about as a
result of the increased storage capabilities, access to data, and the
increased processing power available in the new millennium. These
impressive applications currently infiltrating every industry, are the
result of over 70 years of research and development.
A Brief History of Artificial Intelligence
While computers were first being conceived, we were already wondering how to design them to behave intelligently. In 1946, Alan Turing made the first detailed design of a computer program. In subsequent years, he was intensely concentrated on the problem of artificial intelligence, and created the now-famous test for determining whether a machine can think. The same year the Turing Test was introduced, Iaasac Asimov’s I-Robot featured humanoid robots with artificial intelligence. Asimov, along with other science fiction authors of that era inspired researchers and excited the imaginations of a generation about the potential for artificial intelligence.
During this time, research in neuroscience inspired computer systems called Artificial Neural Networks
(ANN). In an attempt to mirror the way our brains work, the connections
between artificial neurons grow stronger when fired together.
Meanwhile, other researchers were working the idea of creating programs
capable of learning. Ten years after the birth of the computer, the
field of Artificial intelligence was officially born during a workshop
at Dartmouth College, in 1956. A few years later, Arthur Samuel coined the term “Machine Learning”
for the science of using statistics and probability theory to create
algorithms that improve the performance of a given task through
experience rather than explicit instructions. ANN is considered a branch
of ML and is among the earliest techniques in the field of AI.
Cycles of Hype
Since the beginning of AI research, this technology has gone through cycles of hype, leading to over-inflated expectations, disappointment, and loss of funding. In the 50s, the world was introduced to computers playing human games, solving algebra word problems, and learning languages. The intelligence displayed by these machines was simply incredible to the people of the time, and there was much confidence among governments, enterprise, and the academic sphere that these would quickly lead to practical applications. By the 70s, after failing to deliver on the early promise, AI fell into disfavor. Another AI hype cycle began in the 80’s, when a few successful commercial applications brought a renewal of interest. However, it’s application was still not highly developed, and there were difficulties in getting algorithms to do the work expected of them.
The Deep Learning Revolution
In the early 2000’s, neural networks had fallen out of favor, such that it became exceedingly difficult to get research published on the topic. In 2006, a small group of researchers, led by Geoffrey Hinton, made a plan to re-brand neural networks as Deep Learning. Their paper, A fast learning algorithm for deep belief nets, sparked a revival of research into neural networks. It proposed training many more layers than before, with results far exceeding previous attempts. Three years later, Stanford researchers published Large-scale deep unsupervised learning using graphics processors. The GPU became the key to unlocking ANN and other AI techniques. These methods produced results up to 70% faster than any previous attempts, dramatically reducing the time required to perform an experiment. The reason a GPU is so much faster than a CPU is because a CPU must be able to switch very quickly between applications. The GPU on the other hand does not switch between tasks quickly, rather all of it’s power is in repeatedly performing the same type of operation.
In 2009, researchers began working on a dataset to map out the world of images
in a competition known as ImageNet. The next year, deep learning
methods were introduced, dominating the competition. From that point
forward, deep learning drew a lot of attention to AI research, having
its breakout year in 2012. By 2016, deep learning went mainstream and is
now playing a role in our every day lives.
Behind the Scenes
Machine learning occurs in a few stages including data processing, training, deployment and monitoring. First, you must determine what questions you want to ask, or what problems you need to solve. Next, you must determine which data is likely to provide the answer. Then, the data must be gathered and prepared, removing extraneous information, incomplete datasets, and ensuring that it is complete and correctly formatted. Data selection and preparation methods often require running tests on small samples, and transforming the data multiple times.
Once the data is ready, an algorithm must be selected. There are four main categories of ML models: Supervised, Unsupervised, Semi-Supervised, and Reinforcement algorithms.
- Supervised algorithms are used when there is a target output. These models are fed labeled data until they produce the desired results. Those results may include classification of data, prediction, and anomaly detection.
- Unsupervised algorithms are used to find relationships in unlabeled data and can help to eliminate variables.
- Semi-supervised algorithms are fed small amounts of labeled data and taught to classify a large batch of unprocessed data.
- Reinforcement algorithms use trial and error to improve the performance of a software agent within a specific context.
These
models are consistently trained with new data. While seeking an optimal
output, the data may be reformatted multiple times. There are even
algorithms to monitor the progress of primary algorithms, notifying
technicians in case of an anomaly.
Algorithmic Bias
AI influenced applications are quickly changing the enterprise landscape. As ML vastly improves translation and voice recognition apps, it also offers a versatile way to interpret financial data and improve business practices. It’s easy to get excited about the possibilities, however, algorithms are only as good as their designers. Without proper training, machine learning will not produce the desired results. One way that can happen is by using old datasets that don’t accurately reflect the current market. Incomplete data can lead to biased results. We tend to assume that computer code won’t share our biases, but experience has proven otherwise.
Machine
learning is currently used in determining medical care, and making
legal, or financial decisions. Some of these applications have been
found to contain bias. For example, there is a legal product available
that performs a risk assessment of defendants. This program is used to
determine bail and sentences. An independent study found that that this
product labeled African Americans as higher risk, who did not go on to
commit further crime. White defendants, on the other hand, were labeled
as lower risk that did commit new crimes. In 2015, Google’s advertising
engine was proven far less likely to show women advertising for high
paying jobs. These are just two examples of how racism, sexism, and
other types of bias can slip into this far-reaching innovation.
Understandably, there is much concern surrounding the lack of transparency in machine learning practices. AlgorithmWatch,
a non-profit research and advocacy group, was founded to address those
concerns. This group works specifically to evaluate and increase the
transparency of algorithmic decisions with social relevance. They
network with experts from a variety of cultures and disciplines to
explain this complex subject to the general public and develop
strategies to mitigate the effects of bias in algorithmic decision
making. This fledgling industry will require regulation and purposeful
strategies to avoid the danger of bias in the algorithms themselves, and
their training data. Society as a whole needs to be actively engaged
with the values going into automated decision making. That can be a
difficult task, as knowledge about these processes is not widespread and
we only see their results.
Welcome to the Future
While computers aren’t gaining the ability to think, they can
process volumes of information at a rate that’s difficult to
comprehend. That ability has become increasingly important as the modern
world creates more data than it can possibly process. The amount of
data produced is continually growing. Today, one of the most important
tasks in artificial intelligence is in creating algorithms that can
extract value from massive datasets. That fact inspired Harvard Business
Review to declare Data Science as “Sexiest Job in the 21st Century.”
Applications
of artificial intelligence bring an incredible amount of automation to
our lives. For example, these algorithms are able to process an
organization’s entire financial ledger and look for anomalies, allowing
human auditors to examine specific cases that require attention. Google Translate
introduced an AI-driven translation engine that is continually
improving, with user-submitted translations to train and refine the
software. We now have voice-activated digital assistants that are
quickly becoming indispensable, and will soon be standard in the
workplace. Of course you’ve noticed how good social media is at guessing
who you might know, or how Netflix knows what type of movie you might
like to watch.
There are currently applications
that can accurately colorize black and white images, increase the
resolution of low-res photos (like in CSI, but for real), and recognize
the makeup of images. These computer vision applications are already
impressive, and we know they’ll continue to improve. What we don’t know
is how incredible they’ll become. These are just a few examples of the
automation that’s growing more powerful every day.
This technology is creating opportunities in finance, entertainment,
education, science, law, and every other field you can imagine. These
new tools are automating some jobs out of existence. However, they are
expected to create as much work, if not more, than they replace. As
with any major innovation, AI-tech comes with challenges, risks, and
the potential for great rewards. Learning more, and taking advantage of
the opportunities, sooner than later, could give you the edge to get
ahead. Anyone who doesn’t very well may be left behind.
Further Reading
- Why Deep Learning Is Suddenly Changing Your Life
- A ‘Brief’ History of Neural Nets and Deep Learning
- The Revolutionary Technique That Quietly Changed Machine Vision Forever
- Machine Learning Mastery
- The Data Science Process: What a data scientist actually does day-to-day
- www.kdnuggets.com — Machine Learning, Data Science, Data Mining, Big Data, Analytics, AI
- Biased Algorithms Are Everywhere, and No One Seems to Care
- Algorithm Watch
- Google AI Blog: A Neural Network for Machine Translation, at Production Scale
- 100+ AI use cases / applications: A comprehensive guide
No comments:
Post a Comment