Supervised Learning – Dataconomy https://dataconomy.ru Bridging the gap between technology and business Thu, 28 Apr 2022 07:52:57 +0000 en-US hourly 1 https://dataconomy.ru/wp-content/uploads/2025/01/DC_icon-75x75.png Supervised Learning – Dataconomy https://dataconomy.ru 32 32 The history of Machine Learning – dates back to the 17th century https://dataconomy.ru/2022/04/27/the-history-of-machine-learning/ https://dataconomy.ru/2022/04/27/the-history-of-machine-learning/#respond Wed, 27 Apr 2022 15:13:51 +0000 https://dataconomy.ru/?p=23534 Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed, and train themselves in unfamiliar environments, goes back to the 17th century. Machine learning is a powerful tool for implementing artificial intelligence technologies. Because of its ability to learn and make decisions, machine […]]]>

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed, and train themselves in unfamiliar environments, goes back to the 17th century.

Machine learning is a powerful tool for implementing artificial intelligence technologies. Because of its ability to learn and make decisions, machine learning is frequently referred to as AI, even though it is technically a subdivision of AI technology. Until the late 1970s, machine learning was only another component of AI’s progress. It then diverged and evolved on its own, as machine learning has emerged as an important function in cloud computing and e-Commerce. ML is a vital enabler in many cutting-edge technology areas of our times. Scientists are currently working on Quantum Machine Learning approaches.

Remembering the basics

Before embarking on our historical adventure that will span several centuries, let’s briefly go over what we know about Machine Learning (ML).

Today, machine learning is an essential component of business and research for many organizations. It employs algorithms and neural network models to help computers get better at performing tasks. Machine learning algorithms create a mathematical model from data – also known as training data – without being specifically programmed.

The brain cell interaction model that underpins modern machine learning is derived from neuroscience. In 1949, psychologist Donald Hebb published The Organization of Behavior, in which he proposed the idea of “endogenous” or “self-generated” learning. However, it took centuries and crazy inventions like the data-storing weaving loom for us to have such a deep understanding of machine learning as Hebb had in ’49. After this date, other developments in the field were also astonishing and even jaw-dropping on some occasions.

The history of Machine Learning

For ages, we, the people, have been attempting to make sense of data, process it to obtain insights, and automate this process as much as possible. And this is why the technology we now call “machine learning” emerged. Now buckle up, and let’s take on an intriguing journey down the history of machine learning to discover how it all began, how it evolved into what it is today, and what the future may hold for this technology.

· 1642 – The invention of the mechanical adder

Blaise Pascal created one of the first mechanical adding machines as an attempt to automate data processing. It employed a mechanism of cogs and wheels, similar to those in odometers and other counting devices.

Pascal was inspired to build a calculator to assist his father, the superintendent of taxes in Rouen, with the time-consuming arithmetic computations he had to do. He created the device to add and subtract two numbers directly and multiply and divide.

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed
The history of machine learning: Here is a mechanical adder or a basic calculator

The calculator had articulated metal wheel dials with the digits 0 through 9 displayed around the circumference of each wheel. The user inserted a stylus into the corresponding space between the spokes and turned the knob until a metal stop at the bottom was reached to input a digit, similar to how a rotary dial on old phone works. The number is displayed in the top left window of the calculator. Then, simply redialed the second number to be added, resulting in the accumulator’s total being displayed. The carry mechanism, which adds one to nine on one dial and carries one to the next, was another feature of this machine.

· 1801 – The invention of the data storage device

When looking at the history of machine learning, there are lots of surprises. Our first encounter was a data storage device. Believe it or not, the first data storage device was, in fact, a weaving loom. The first use of data storage was in a loom created by a French inventor named Joseph-Marie Jacquard, that used metal cards with holes to arrange threads. These cards comprised a program to control the loom and allowed a procedure to be repeated with the same outcome every time.

The history of Machine Learning - dates back to the 17th century
The history of Machine Learning: A Jacquard loom showing information punchcards, National Museum of Scotland

The Jacquard Machine used interchangeable punched cards to weave the cloth in any pattern without human intervention. The punched cards were used by Charles Babbage, the famous English inventor, as an input-output medium for his theoretical, analytical engine and by Herman Hollerith to feed data to his census machine. They were also utilized to input data into digital computers, but they have been superseded by electronic equipment.

· 1847 – The introduction of Boolean Logic

In Boolean Logic (also known as Boolean Algebra), all values are either True or False. These true and false values are employed to check the conditions that selection and iteration rely on. This is how Boolean operators work. George Boole created AND, OR, and NOR operators using this logic, responding to questions about true or false, yes or no, and binary 1s and 0s. These operators are still used in web searches today.

Boolean algebra is introduced in artificial intelligence to address some of the problems associated with machine learning. One of the main disadvantages of this discipline is that machine-learning algorithms are black boxes, which means we don’t know a lot about how they autonomously operate. Random forest and decision trees are examples of machine learning algorithms that can describe the functioning of a system, but they don’t always provide excellent results. Boolean algebra is used to overcome this limitation. Boolean algebra has been used in machine learning to produce sets of understandable rules that can achieve quite good performance.

After reading the history of machine learning, you might want to check out 75 Big Data terms everyone should know.

· 1890 – The Hollerith Machine took on statistical calculations

Herman Hollerith developed the first combined mechanical calculation and punch-card system to compute statistics from millions of individuals efficiently. It was an electromechanical machine built to assist in summarizing data stored on punched cards.

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed
The history of machine learning: Statistical calculations were first made with electromechanical machines

The 1890 census in the United States took eight years to complete. Because the Constitution requires a census every ten years, a larger workforce was necessary to expedite the process. The tabulating machine was created to aid in processing 1890 Census data. Later versions were widely used in commercial accounting and inventory management applications. It gave rise to a class of machines known as unit record equipment and the data processing industry.

· 1943 – The first mathematical model of a biological neuron presented

The scientific article “A Logical Calculus of the Ideas Immanent in Nervous Activity,” published by Walter Pitts and Warren McCulloch, introduced the first mathematical model of neural networks. For many, that paper was the real starting point for the modern discipline of machine learning, which led the way for deep learning and quantum machine learning.

McCulloch and Pitts’s 1948 paper built on Alan Turing’s “On Computable Numbers” to provide a means for describing brain activities in general terms, demonstrating that basic components linked in a neural network might have enormous computational capability. Until the ideas were applied by John von Neuman, the architect of modern computing, Norbert Wiene, and others, the paper received little attention.

· 1949 – Hebb successfully related behavior to neural networks and brain activity

In 1949, Canadian psychologist Donald O. Hebb, then a lecturer at McGill University, published The Organization of Behavior: A Neuropsychological Theory. This was the first time that a physiological learning rule for synaptic change had been made explicit in print and became known as the “Hebb synapse.” 

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed
The history of machine learning: Neural networks are used in many AI systems today

McCulloch and Pitts developed cell assembly theory in their 1951 paper. McCulloch and Pitts’ model was later known as Hebbian theory, Hebb’s rule, Hebb’s postulate, and cell assembly theory. Models that follow this idea are said to exhibit “Hebbian learning.” As stated in the book: “When an axon of cell A is near enough to excite cell B and repeatedly or persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A’s efficiency, as one of the cells firing B, is increased.”

Hebb’s model paved the way for the development of computational machines that replicated natural neurological processes

Hebb referred to the combination of neurons that may be regarded as a single processing unit as “cell assemblies.” And their connection mix determined the brain’s change in response to stimuli.

Hebb’s model for the functioning of the mind has had a significant influence on how psychologists view stimulus processing in mind. It also paved the way for the development of computational machines that replicated natural neurological processes, such as machine learning. While chemical transmission became the major form of synaptic transmission in the nervous system, modern artificial neural networks are still built on the foundation of electrical signals traveling through wires that Hebbian theory was created around.

·  1950 – Turing found a way to measure the thinking capabilities of machines

The Turing Test is a test of artificial intelligence (AI) for determining whether or not a computer thinks like a human. The term “Turing Test” derives from Alan Turing, an English computer scientist, cryptanalyst, mathematician, and theoretical biologist who invented the test.

It is impossible to define intelligence in a machine, according to Turing. If a computer can mimic human responses under specific circumstances, it may be said to have artificial intelligence. The original Turing Test requires three physically separated terminals from one another. One terminal is controlled by a computer, while humans use the other two.

The history of Machine Learning - dates back to the 17th century
The history of Machine Learning: The IBM 700 series made scientific calculations and commercial operations easier, but the machines also provided the world with some entertainment (Image courtesy of IBM)

During the experiment, one of the humans serves as the questioner, with the second human and computer as respondents. The questioner asks questions of the respondents in a specific area of study within a specified format and context. After a determined duration or number of queries, the questioner is invited to select which respondent was real and which was artificial. The test is carried out numerous times. The computer is called “artificial intelligence” if the inquirer confirms the correct outcome in half of the test runs or fewer.

The test was named after Alan Turing, who pioneered machine learning during the 40s and 50s. In 1950, Turing published a “Computing Machinery and Intelligence” paper to outline the test.

· 1952 – The first computer learning program was developed at IBM

Arthur Samuel’s Checkers program, which was created for play on the IBM 701, was shown to the public for the first time on television on February 24, 1956. Robert Nealey, a self-described checkers master, played the game on an IBM 7094 computer in 1962. The computer won. The Samuel Checkers program lost other games to Nealey. However, it was still regarded as a milestone for artificial intelligence and provided the public with an example of the abilities of an electronic computer in the early 1960s.

The more the program played, learning which moves made up winning strategies in a ‘supervised learning mode,’ and incorporating them into its algorithm, the better it performed at the game.

Samuel’s program was a groundbreaking story for the time. Computers could beat checkers for the first time. Electronic creations were challenging humanity’s intellectual advantage. To the technology-illiterate public of 1962, this was a significant event. It established the groundwork for machines to do other intelligent tasks better than humans. And people started to think; will computers surpass humans in intelligence? After all, computers were only around for a few years back then, and the artificial intelligence field was still in its infancy…

Moving on in the history of machine learning, you might also want to check out Machine learning engineering: The science of building reliable AI systems.

· 1958 – The Perceptron was designed

In July 1958, the United States Office of Naval Research unveiled a remarkable invention: The perception. An IBM 704 – a 5-ton computer size of a room, was fed a series of punch cards and, after 50 tries, learned to identify cards with markings on the left from markings on the right.

According to its inventor, Frank Rosenblatt, it was a show of the “perceptron,” which was “the first machine capable of generating an original thought,” according to its inventor, Frank Rosenblatt.

“Stories about the creation of machines having human qualities have long been a fascinating province in the realm of science fiction,” Rosenblatt observed in 1958. “Yet we are about to witness the birth of such a machine – a machine capable of perceiving, recognizing, and identifying its surroundings without any human training or control.”

He was right about his vision, but it took almost half a decade to provide it.

· The 60s – Bell Labs’ attempt to teach machines how to read

The term “deep learning” was inspired by a report from the late 1960s describing how scientists at Bell Labs were attempting to teach computers to read English text. The invention of artificial intelligence, or “AI,” in the early 1950s began the trend toward what is now known as machine learning.

· 1967 – Machines gained the ability to recognize patterns 

The “nearest neighbor” algorithm was created, allowing computers to conduct rudimentary pattern detection. When the program was given a new object, it compared it to the existing data and classified it as the nearest neighbor, which meant the most similar item in memory.

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed
The history of machine learning: Pattern recognition is the basis of many AI developments achieved till now

The invention of the pattern recognition algorithm is credited to Fix and Hodges, who detailed their non-parametric technique for pattern classification in 1951 in an unpublished issue of a US Air Force School of Aviation Medicine report. The k-nearest neighbor rule was initially introduced by Fix and Hodges as a non-parametric method for pattern classification.

· 1979 – One of the first autonomous vehicles was invented at Stanford

The Stanford Cart was a decades-long endeavor that evolved in various forms from 1960 to 1980. It began as a study of what it would be like to operate a lunar rover from Earth and was eventually revitalized as an autonomous vehicle. On its own, the student invention cart could maneuver around obstacles in a room. The Stanford Cart was initially a remote-controlled television-equipped mobile robot.

The history of Machine Learning - dates back to the 17th century
The history of Machine Learning: The infamous Stanford Cart (Image courtesy of Stanford University)

A computer program was created to control the Cart through chaotic locations, obtaining all of its information about the world from on-board TV images. The Cart used a variety of stereopsis to discover things in three dimensions and determine its own motion. Based on a model created with this data, it planned an obstacle-avoiding route to the target destination. As the Cart encountered new obstacles on its trip, the plan evolved.

We are talking about the history of machine learning, but data science is also advanced today in many areas. Here are a couple interesting articles we prepared before:

· 1981 – Explanation based learning prompt to supervised learning

Gerald Dejong pioneered explanation-based learning (EBL) in a journal article published in 1981. EBL laid the foundation of modern supervised learning because training examples supplement prior knowledge of the world. The program analyzes the training data and eliminates unneeded information to create a broad rule applied to future instances. For example, if the software is instructed to concentrate on the queen in chess, it will discard all non-immediate-effect pieces.

· The 90s – Emergence of various machine learning applications 

Scientists began to apply machine learning in data mining, adaptive software, web applications, text learning, and language learning in the 1990s. Scientists create computer programs that can analyze massive amounts of data and draw conclusions or learn from the findings. The term “Machine Learning” was coined as scientists were finally able to develop software in such a way that it could learn and improve on its own, requiring no human input.

· The Millennium – The rise of adaptive programming

The new millennium saw an unprecedented boom in adaptive programming. Machine learning went hand to hand with adaptive solutions for a long time. These programs can identify patterns, learn from experience, and improve themselves based on the feedback they receive from the environment.

Deep learning is an example of adaptive programming, where algorithms can “see” and distinguish objects in pictures and videos, which was the underlying technology behind Amazon GO shops. Customers are charged as they walk out without having to stand in line.

amazon-go
The history of Machine Learning: Amazon GO shops charge customers as they walk out without standing in line (Image courtesy of Amazon)

· Today – Machine learning is a valuable tool for all industries

Machine learning is one of today’s cutting-edge technologies that has aided us in improving not just industrial and professional procedures but also day-to-day life. This branch of machine learning uses statistical methods to create intelligent computer systems capable of learning from data sources accessible to it.

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed
The history of machine learning: Medical diagnosis is one area that ML will change soon

Machine learning is already being utilized in various areas and sectors. Medical diagnosis, image processing, prediction, classification, learning association, and regression are just a few applications. Machine learning algorithms are capable of learning from previous experiences or historical data. Machine learning programs use the experience to produce outcomes.

Organizations use machine learning to gain insight into consumer trends and operational patterns, as well as the creation of new products. Many of today’s top businesses incorporate machine learning into their daily operations. For many businesses, machine learning has become a significant competitive differentiator. In fact, machine learning engineering is a rising area.

· Tomorrow – The future of Machine Learning: Chasing the quantum advantage

Actually, our article was supposed to end here, since we came to today in the history of machine learning, but it doesn’t, because tomorrow holds more…

For example, Quantum Machine Learning (QML) is a young theoretical field investigating the interaction between quantum computing and machine learning methods. Quantum computing has recently been shown to have advantages for machine learning in several experiments. The overall objective of Quantum Machine Learning is to make things move faster by combining what we know about quantum computing with conventional machine learning. The idea of Quantum Machine Learning is derived from classical Machine Learning theory and interpreted in that light.

The application of quantum computers in the real world has advanced rapidly during the last decade, with the potential benefit becoming more apparent. One important area of research is how quantum computers may affect machine learning. It’s recently been demonstrated experimentally that quantum computers can solve problems with complex correlations between inputs that are difficult for traditional systems.

According to Google’s research, quantum computers may be more beneficial in certain applications. Quantum models generated on quantum computing machines might be far more potent for particular tasks, allowing for quicker processing and generalization on fewer data. As a result, it’s crucial to figure out when such a quantum edge can be exploited…

]]>
https://dataconomy.ru/2022/04/27/the-history-of-machine-learning/feed/ 0
How Faulty Data Breaks Your Machine Learning Process https://dataconomy.ru/2017/06/23/faulty-data-machine-learning/ https://dataconomy.ru/2017/06/23/faulty-data-machine-learning/#respond Fri, 23 Jun 2017 09:00:03 +0000 https://dataconomy.ru/?p=18143 This article is part of a media partnership with PyData Berlin, a group helping support open-source data science libraries and tools. To learn more about this topic, please consider attending our fourth annual PyData Berlin conference on June 30-July 2, 2017. Miroslav Batchkarov and other experts will be giving talks on Natural Language Processing, Machine […]]]>

How Faulty Data Breaks Your Machine Learning Process
This article is part of a media partnership with PyData Berlin, a group helping support open-source data science libraries and tools. To learn more about this topic, please consider attending our fourth annual
PyData Berlin conference
on June 30-July 2, 2017. Miroslav Batchkarov and other experts will be giving talks on Natural Language Processing, Machine Learning, AI Ethics and many related data fields. You can also find a more detailed blog post on Miroslav Batchkarov’s personal blog at https://mbatchkarov.github.io


Introduction

It is often said that rather than spending a month figuring out how to apply unsupervised learning to a problem domain, a data scientist should spend a week labeling data. However, the difficulty of annotating data is often underestimated. Gathering a sufficiently large collection of good-quality labeled data requires careful problem definition, quality control and multiple iterations. As a result, gathering enough data to build a high-accuracy supervised model can take much longer than one might expect. This post describes my experiences in labeling gold-standard data for natural language processing and the lessons learned along the way.

Case study 1: word embeddings

Words vectors have become popular in recent years as they can be built without supervision and can capture complex semantic relations well. For example, adding the vectors of king and woman and subtracting the vector of man yields a vector that is very close to that of queen.

Algorithms for training word vectors have been evaluated by correlating the “similarity” of word pairs, as predicted by a model, to those provided by a human judge. A typical data set consists of word pairs and a similarity score, e.g. cat, dog, 80% and cat, purple, 21%. A model is considered good if it assigns high scores to word pairs that are scored highly by a human. Let us consider what it takes to label such a data set and what can go wrong in the process.

Is the task clearly defined?

What exactly is word similarity? We all have an intuitive understanding, but a lot of corner cases are hard to pin down. What makes the words cat and dog similar? Is it because they are both animals? What about topically related words, such as rice and cooking? What about antonyms, e.g. big and small? Human annotators need clear and unambiguous description of what is required of them, otherwise data quality will be
poor. For instance, the similarity scores provided by 13 annotators for the pair tiger–cat range from 50% to 90% in WordSim353, a data set often used to rank word similarity. Click here for a more detailed analysis of this issue.

Is the task easy for humans to do?

Even with clear instructions, some tasks are inherently subjective. It is unlikely that every person you ask will provide the same similarity score for cat and dog, or will interpret a written sarcastic comment the same way out of context. If humans cannot agree what the right answer is for a given input, how can a model ever do well? Of course, data scientists can take steps to address those issues. First, make sure that you have clear written annotator guidelines, even if you plan to do the annotation yourself. Ask others to read the guidelines and explain how they interpret your instructions.

Second, do not be afraid to change the task to make it easier. If your use case allows it, make the task as simple as possible for the annotators. Ask yourself if it is business-critical to get fine-grained labels or if a coarser-grained (and therefore easier) annotation schema would suffice.

Do you have quality control in place?

Not having a mechanism for identifying annotator errors or consistently under-performing annotators is perhaps the most common error in data labeling. There are many viable ways to do quality control – the key is to ensure at least one of them is in use. A common approach is to measure if different annotators agree with one another or with a known gold-standard – see next case study. Conflicts may be resolved by an independent adjudicator.

Ideally, quality control should be an automated process that runs continuously. Remember that systematic errors may be a sign of unclear guidelines. Talk to your annotators to understand the source of the problem. Be prepared to discard your first data set.

How much data can you hope to get?

A supervised learning model may require a significant amount of labeled data to perform well. However, obtaining this much data may be problematic for a number of reasons:

  • Human annotators are expensive, even on crowd sourcing platforms that systematically underpay.
  • Good, trustworthy annotators are hard to find, especially if they need special training.
    • An aside on ethics: pay fairly and treat annotators with respect at all times. The less you pay, the higher the odds you will have to repeat the job.
  • Errors in labeling may be hard to correct, so be prepared to discard data.

Keep an eye on the learning curve of your model. Do not rush into gathering more data unless there is evidence to suggest it would help. It is often better to focus on quality rather than quantity.

Case study 2: symptom recognition in medical data

The second case study involves identifying mentions of symptoms or diseases in notes taken by a doctor during routine exams. An indeal note may read: Abdominal pain due to acute bacterial tonsilitis. No allergies. However, these notes are often taken in a hurry and typically look more like this:

Abd pian//acute b tons//n/a alergies.

This presents a different set of challenges for a data scientist.

Do you need an expert annotator?

Most doctors’ notes are impossible to decode for a layperson. Annotation therefore has to be done by a trained doctor, but most doctors prefer to practice medicine rather than label data. The typical doctor is also not an expert in machine learning or linguistics (and vice versa), so they may not have the same vocabulary. This makes it harder to provide a clear task definition.

Can you measure inter-annotator agreement?

Can you quantify the degree to which annotators agree? In the case of word similarity, this is as simple as comparing numerical scores. This gets trickier when the annotation unit has complex structure (e.g. it is a phrase). For example, in the sentence burning neck pain one annotator may pick up the first two words as a symptom, while another may pick up the last two words, and a third may not mark any symptoms.

Tooling

Do you need specialist software to capture and store data? Depending on the complexity, can you write something yourself or do you have to assemble a team? Will the tool be intuitive and easy to use, or is it confusing your annotators? In our medical example, the tooling is getting much better and allows annotators to be very productive. Several years ago tools such as BRAT were less mature and took a long time to set up, whereas now they run out of the box.

Further issues arise where hardware is involved, e.g. in internet-of-things or industrial applications. These include writing bespoke software for the device to capture data, transmitting it to a server reliably (and securely!) and storing potentially massive amount of data that machines generate.

People issues

Many data science talks focus on the technical aspects of data gathering, with people issues often taking a back seat. However, a single well-trained and reliable annotator producing large amounts of labeled data can often contribute more to the success of a project than a team of programmers. It is therefore important to build and maintain a good working relationship with the annotators. This is much easier if they are located near you. Be prepared to deal with small issues such as
annotators not showing up for sessions, or taking two-week breaks between sessions. The more time passes between sessions, the more the annotators will need re-training. This is the main reason why quality control should be run continuously.

If you are using a crowd sourcing platform, beware of click farms. The academic literature has some fantastic resources on the subject- see references.

General lessons

  • Get to know the problem domain
  • Do not be afraid to start from scratch if your assumptions are wrong
  • Monitor quality continuously
  • Beware of crowd-sourcing

References and notes

The first case study is based on this paper, which is in turn based on my PhD work. The second case study is inspired by the PhD work of my lab mate Aleksandar Savkov.

  • Herbert Rubenstein and John Goodenough. 1965. Contextual correlates of synonymy. Communications of the ACM 8(10):627-633.
  • Felix Hill, Roi Reichart, and Anna Korhonen. 2015. Simlex-999: Evaluating semantic models with (genuine) similarity estimation. Computational Linguistics
  • Lev Finkelstein, Evgeniy Gabrilovich, Yossi Matias, Ehud Rivlin, Zach Solan, Gadi Wolfman, and Eytan Ruppin. 2001. Placing search in context: The concept revisited. Proceedings of the 10th international conference on World Wide Web pages 406-414.
  • George Miller and Walter Charles. 1991. Contextual correlates of semantic similarity. Language and Cognitive Processes 6(1):1-28.
  • Chris Bieman. Crowdsourcing Linguistic Datasets
  • Chris Carlison-Burch et al. Crowdsourcing and Human Computation Class
  • Rion Snow, Brendan O’Connor, Daniel Jurafsky, and Andrew Y. Ng. 2008. Cheap and fast- but is it good?: evaluating non-expert annotations for natural language tasks.
  • Gadiraju, U., Kawase, R., Dietze, S., Demartini, G. 2015. Understanding Malicious Behaviour in Crowdsourcing Platforms: The Case of Online Surveys.

Like this article? Subscribe to our weekly newsletter to never miss out!

]]>
https://dataconomy.ru/2017/06/23/faulty-data-machine-learning/feed/ 0
What’s The Difference Between Supervised and Unsupervised Learning? https://dataconomy.ru/2015/01/08/whats-the-difference-between-supervised-and-unsupervised-learning/ https://dataconomy.ru/2015/01/08/whats-the-difference-between-supervised-and-unsupervised-learning/#comments Thu, 08 Jan 2015 15:43:27 +0000 https://dataconomy.ru/?p=11299 Wiki Supervised Learning Definition  Supervised learning is the Data mining task of inferring a function from labeled training data.The training data consist of a set of training examples. In supervised learning, each example is a pair consisting of an input object (typically a vector) and a desired output value (also called thesupervisory signal). A supervised learning […]]]>

Wiki Supervised Learning Definition 

Supervised learning is the Data mining task of inferring a function from labeled training data.The training data consist of a set of training examples. In supervised learning, each example is a pair consisting of an input object (typically a vector) and a desired output value (also called thesupervisory signal). A supervised learning algorithm analyzes the training data and produces an inferred function, which can be used for mapping new examples. An optimal scenario will allow for the algorithm to correctly determine the class labels for unseen instances. This requires the learning algorithm to generalize from the training data to unseen situations in a “reasonable” way.

Wiki Unsupervised Learning Definition

In Data mining, the problem of unsupervised learning is that of trying to find hidden structure in unlabeled data. Since the examples given to the learner are unlabeled, there is no error or reward signal to evaluate a potential solution.

______

On November 25th-26th 2019, we are bringing together a global community of data-driven pioneers to talk about the latest trends in tech & data at Data Natives Conference 2019. Get your ticket now at a discounted Early Bird price!

_____

Let’s learn supervised and unsupervised learning with a real life example

[bctt tweet=”Learn supervised and unsupervised learning with a real life example:”]

basket

    • suppose you had a basket and it is fulled with some different kinds of fruits, your task is to arrange them as groups.
    • For understanding let me clear the names of the fruits in our basket.
    • We have four types of fruits. They are: apple, banana, grape and cherry.

Supervised Learning :

  • You already learn from your previous work about the physical characters of fruits.
  • So arranging  the same type of fruits at one place is easy now.
  • Your previous work is called as training data in data mining.
  • so you already learn the things from your train data, this is because of response variable.
  • Response variable mean just a decision variable.
  • You can observe response variable below (FRUIT NAME) .
NO. SIZE COLOR SHAPE FRUIT NAME
1 Big Red Rounded shape with a depression at the top Apple
2 Small Red Heart-shaped to nearly globular Cherry
3 Big Green Long curving cylinder Banana
4 Small Green Round to oval,Bunch shape Cylindrical Grape
  • Suppose you have  taken an new fruit from the basket then you will see the size , color and shape of that particular fruit.
  • If  size  is Big , color is Red , shape is rounded shape with a depression at the top, you will conform the fruit name as apple and you will put in apple group.
  • Likewise for other fruits also.
  • Job of groping fruits was done and happy ending.
  • You can observe in the table that  a column was labeled as “FRUIT NAME” this is called as response variable.
  • If you learn the thing before from training data and then applying that knowledge to the test data(for new fruit), This type of learning is called as Supervised Learning.
  • Classification come under Supervised learning.

Unsupervised Learning

  • Suppose you had a basket and it is fulled with some different types fruits, your task is to arrange them as groups.
  • This time you don’t know any thing about that fruits, honestly saying this is the first time you have seen them.
  • so how will you arrange them.
  • What will you do first???
  • You will take a fruit and you will arrange them by considering physical character of that particular fruit. suppose you have considered color.
  • Then you will arrange them on considering base condition as color.
  • Then the groups will be some thing like this.
  • RED COLOR GROUP: apples & cherry fruits.
  • GREEN COLOR GROUP: bananas & grapes.
  • so now you will take another physical character such as size .
  • RED COLOR AND BIG SIZE: apple.
  • RED COLOR AND SMALL SIZE: cherry fruits.
  • GREEN COLOR AND BIG SIZE: bananas.
  • GREEN COLOR AND SMALL SIZE: grapes.
  • job done happy ending.
  • Here you didn’t know learn any thing before ,means no train data and no response variable.
  • This type of learning is know unsupervised learning.
  • clustering comes under unsupervised learning.

George-Clooney5

This post originally appeared here.


(Image credit: Wata1219, via Flickr)

]]>
https://dataconomy.ru/2015/01/08/whats-the-difference-between-supervised-and-unsupervised-learning/feed/ 9