Timeline – Dataconomy https://dataconomy.ru Bridging the gap between technology and business Thu, 28 Apr 2022 07:52:57 +0000 en-US hourly 1 https://dataconomy.ru/wp-content/uploads/2025/01/DC_icon-75x75.png Timeline – Dataconomy https://dataconomy.ru 32 32 The history of Machine Learning – dates back to the 17th century https://dataconomy.ru/2022/04/27/the-history-of-machine-learning/ https://dataconomy.ru/2022/04/27/the-history-of-machine-learning/#respond Wed, 27 Apr 2022 15:13:51 +0000 https://dataconomy.ru/?p=23534 Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed, and train themselves in unfamiliar environments, goes back to the 17th century. Machine learning is a powerful tool for implementing artificial intelligence technologies. Because of its ability to learn and make decisions, machine […]]]>

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed, and train themselves in unfamiliar environments, goes back to the 17th century.

Machine learning is a powerful tool for implementing artificial intelligence technologies. Because of its ability to learn and make decisions, machine learning is frequently referred to as AI, even though it is technically a subdivision of AI technology. Until the late 1970s, machine learning was only another component of AI’s progress. It then diverged and evolved on its own, as machine learning has emerged as an important function in cloud computing and e-Commerce. ML is a vital enabler in many cutting-edge technology areas of our times. Scientists are currently working on Quantum Machine Learning approaches.

Remembering the basics

Before embarking on our historical adventure that will span several centuries, let’s briefly go over what we know about Machine Learning (ML).

Today, machine learning is an essential component of business and research for many organizations. It employs algorithms and neural network models to help computers get better at performing tasks. Machine learning algorithms create a mathematical model from data – also known as training data – without being specifically programmed.

The brain cell interaction model that underpins modern machine learning is derived from neuroscience. In 1949, psychologist Donald Hebb published The Organization of Behavior, in which he proposed the idea of “endogenous” or “self-generated” learning. However, it took centuries and crazy inventions like the data-storing weaving loom for us to have such a deep understanding of machine learning as Hebb had in ’49. After this date, other developments in the field were also astonishing and even jaw-dropping on some occasions.

The history of Machine Learning

For ages, we, the people, have been attempting to make sense of data, process it to obtain insights, and automate this process as much as possible. And this is why the technology we now call “machine learning” emerged. Now buckle up, and let’s take on an intriguing journey down the history of machine learning to discover how it all began, how it evolved into what it is today, and what the future may hold for this technology.

· 1642 – The invention of the mechanical adder

Blaise Pascal created one of the first mechanical adding machines as an attempt to automate data processing. It employed a mechanism of cogs and wheels, similar to those in odometers and other counting devices.

Pascal was inspired to build a calculator to assist his father, the superintendent of taxes in Rouen, with the time-consuming arithmetic computations he had to do. He created the device to add and subtract two numbers directly and multiply and divide.

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed
The history of machine learning: Here is a mechanical adder or a basic calculator

The calculator had articulated metal wheel dials with the digits 0 through 9 displayed around the circumference of each wheel. The user inserted a stylus into the corresponding space between the spokes and turned the knob until a metal stop at the bottom was reached to input a digit, similar to how a rotary dial on old phone works. The number is displayed in the top left window of the calculator. Then, simply redialed the second number to be added, resulting in the accumulator’s total being displayed. The carry mechanism, which adds one to nine on one dial and carries one to the next, was another feature of this machine.

· 1801 – The invention of the data storage device

When looking at the history of machine learning, there are lots of surprises. Our first encounter was a data storage device. Believe it or not, the first data storage device was, in fact, a weaving loom. The first use of data storage was in a loom created by a French inventor named Joseph-Marie Jacquard, that used metal cards with holes to arrange threads. These cards comprised a program to control the loom and allowed a procedure to be repeated with the same outcome every time.

The history of Machine Learning - dates back to the 17th century
The history of Machine Learning: A Jacquard loom showing information punchcards, National Museum of Scotland

The Jacquard Machine used interchangeable punched cards to weave the cloth in any pattern without human intervention. The punched cards were used by Charles Babbage, the famous English inventor, as an input-output medium for his theoretical, analytical engine and by Herman Hollerith to feed data to his census machine. They were also utilized to input data into digital computers, but they have been superseded by electronic equipment.

· 1847 – The introduction of Boolean Logic

In Boolean Logic (also known as Boolean Algebra), all values are either True or False. These true and false values are employed to check the conditions that selection and iteration rely on. This is how Boolean operators work. George Boole created AND, OR, and NOR operators using this logic, responding to questions about true or false, yes or no, and binary 1s and 0s. These operators are still used in web searches today.

Boolean algebra is introduced in artificial intelligence to address some of the problems associated with machine learning. One of the main disadvantages of this discipline is that machine-learning algorithms are black boxes, which means we don’t know a lot about how they autonomously operate. Random forest and decision trees are examples of machine learning algorithms that can describe the functioning of a system, but they don’t always provide excellent results. Boolean algebra is used to overcome this limitation. Boolean algebra has been used in machine learning to produce sets of understandable rules that can achieve quite good performance.

After reading the history of machine learning, you might want to check out 75 Big Data terms everyone should know.

· 1890 – The Hollerith Machine took on statistical calculations

Herman Hollerith developed the first combined mechanical calculation and punch-card system to compute statistics from millions of individuals efficiently. It was an electromechanical machine built to assist in summarizing data stored on punched cards.

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed
The history of machine learning: Statistical calculations were first made with electromechanical machines

The 1890 census in the United States took eight years to complete. Because the Constitution requires a census every ten years, a larger workforce was necessary to expedite the process. The tabulating machine was created to aid in processing 1890 Census data. Later versions were widely used in commercial accounting and inventory management applications. It gave rise to a class of machines known as unit record equipment and the data processing industry.

· 1943 – The first mathematical model of a biological neuron presented

The scientific article “A Logical Calculus of the Ideas Immanent in Nervous Activity,” published by Walter Pitts and Warren McCulloch, introduced the first mathematical model of neural networks. For many, that paper was the real starting point for the modern discipline of machine learning, which led the way for deep learning and quantum machine learning.

McCulloch and Pitts’s 1948 paper built on Alan Turing’s “On Computable Numbers” to provide a means for describing brain activities in general terms, demonstrating that basic components linked in a neural network might have enormous computational capability. Until the ideas were applied by John von Neuman, the architect of modern computing, Norbert Wiene, and others, the paper received little attention.

· 1949 – Hebb successfully related behavior to neural networks and brain activity

In 1949, Canadian psychologist Donald O. Hebb, then a lecturer at McGill University, published The Organization of Behavior: A Neuropsychological Theory. This was the first time that a physiological learning rule for synaptic change had been made explicit in print and became known as the “Hebb synapse.” 

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed
The history of machine learning: Neural networks are used in many AI systems today

McCulloch and Pitts developed cell assembly theory in their 1951 paper. McCulloch and Pitts’ model was later known as Hebbian theory, Hebb’s rule, Hebb’s postulate, and cell assembly theory. Models that follow this idea are said to exhibit “Hebbian learning.” As stated in the book: “When an axon of cell A is near enough to excite cell B and repeatedly or persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A’s efficiency, as one of the cells firing B, is increased.”

Hebb’s model paved the way for the development of computational machines that replicated natural neurological processes

Hebb referred to the combination of neurons that may be regarded as a single processing unit as “cell assemblies.” And their connection mix determined the brain’s change in response to stimuli.

Hebb’s model for the functioning of the mind has had a significant influence on how psychologists view stimulus processing in mind. It also paved the way for the development of computational machines that replicated natural neurological processes, such as machine learning. While chemical transmission became the major form of synaptic transmission in the nervous system, modern artificial neural networks are still built on the foundation of electrical signals traveling through wires that Hebbian theory was created around.

·  1950 – Turing found a way to measure the thinking capabilities of machines

The Turing Test is a test of artificial intelligence (AI) for determining whether or not a computer thinks like a human. The term “Turing Test” derives from Alan Turing, an English computer scientist, cryptanalyst, mathematician, and theoretical biologist who invented the test.

It is impossible to define intelligence in a machine, according to Turing. If a computer can mimic human responses under specific circumstances, it may be said to have artificial intelligence. The original Turing Test requires three physically separated terminals from one another. One terminal is controlled by a computer, while humans use the other two.

The history of Machine Learning - dates back to the 17th century
The history of Machine Learning: The IBM 700 series made scientific calculations and commercial operations easier, but the machines also provided the world with some entertainment (Image courtesy of IBM)

During the experiment, one of the humans serves as the questioner, with the second human and computer as respondents. The questioner asks questions of the respondents in a specific area of study within a specified format and context. After a determined duration or number of queries, the questioner is invited to select which respondent was real and which was artificial. The test is carried out numerous times. The computer is called “artificial intelligence” if the inquirer confirms the correct outcome in half of the test runs or fewer.

The test was named after Alan Turing, who pioneered machine learning during the 40s and 50s. In 1950, Turing published a “Computing Machinery and Intelligence” paper to outline the test.

· 1952 – The first computer learning program was developed at IBM

Arthur Samuel’s Checkers program, which was created for play on the IBM 701, was shown to the public for the first time on television on February 24, 1956. Robert Nealey, a self-described checkers master, played the game on an IBM 7094 computer in 1962. The computer won. The Samuel Checkers program lost other games to Nealey. However, it was still regarded as a milestone for artificial intelligence and provided the public with an example of the abilities of an electronic computer in the early 1960s.

The more the program played, learning which moves made up winning strategies in a ‘supervised learning mode,’ and incorporating them into its algorithm, the better it performed at the game.

Samuel’s program was a groundbreaking story for the time. Computers could beat checkers for the first time. Electronic creations were challenging humanity’s intellectual advantage. To the technology-illiterate public of 1962, this was a significant event. It established the groundwork for machines to do other intelligent tasks better than humans. And people started to think; will computers surpass humans in intelligence? After all, computers were only around for a few years back then, and the artificial intelligence field was still in its infancy…

Moving on in the history of machine learning, you might also want to check out Machine learning engineering: The science of building reliable AI systems.

· 1958 – The Perceptron was designed

In July 1958, the United States Office of Naval Research unveiled a remarkable invention: The perception. An IBM 704 – a 5-ton computer size of a room, was fed a series of punch cards and, after 50 tries, learned to identify cards with markings on the left from markings on the right.

According to its inventor, Frank Rosenblatt, it was a show of the “perceptron,” which was “the first machine capable of generating an original thought,” according to its inventor, Frank Rosenblatt.

“Stories about the creation of machines having human qualities have long been a fascinating province in the realm of science fiction,” Rosenblatt observed in 1958. “Yet we are about to witness the birth of such a machine – a machine capable of perceiving, recognizing, and identifying its surroundings without any human training or control.”

He was right about his vision, but it took almost half a decade to provide it.

· The 60s – Bell Labs’ attempt to teach machines how to read

The term “deep learning” was inspired by a report from the late 1960s describing how scientists at Bell Labs were attempting to teach computers to read English text. The invention of artificial intelligence, or “AI,” in the early 1950s began the trend toward what is now known as machine learning.

· 1967 – Machines gained the ability to recognize patterns 

The “nearest neighbor” algorithm was created, allowing computers to conduct rudimentary pattern detection. When the program was given a new object, it compared it to the existing data and classified it as the nearest neighbor, which meant the most similar item in memory.

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed
The history of machine learning: Pattern recognition is the basis of many AI developments achieved till now

The invention of the pattern recognition algorithm is credited to Fix and Hodges, who detailed their non-parametric technique for pattern classification in 1951 in an unpublished issue of a US Air Force School of Aviation Medicine report. The k-nearest neighbor rule was initially introduced by Fix and Hodges as a non-parametric method for pattern classification.

· 1979 – One of the first autonomous vehicles was invented at Stanford

The Stanford Cart was a decades-long endeavor that evolved in various forms from 1960 to 1980. It began as a study of what it would be like to operate a lunar rover from Earth and was eventually revitalized as an autonomous vehicle. On its own, the student invention cart could maneuver around obstacles in a room. The Stanford Cart was initially a remote-controlled television-equipped mobile robot.

The history of Machine Learning - dates back to the 17th century
The history of Machine Learning: The infamous Stanford Cart (Image courtesy of Stanford University)

A computer program was created to control the Cart through chaotic locations, obtaining all of its information about the world from on-board TV images. The Cart used a variety of stereopsis to discover things in three dimensions and determine its own motion. Based on a model created with this data, it planned an obstacle-avoiding route to the target destination. As the Cart encountered new obstacles on its trip, the plan evolved.

We are talking about the history of machine learning, but data science is also advanced today in many areas. Here are a couple interesting articles we prepared before:

· 1981 – Explanation based learning prompt to supervised learning

Gerald Dejong pioneered explanation-based learning (EBL) in a journal article published in 1981. EBL laid the foundation of modern supervised learning because training examples supplement prior knowledge of the world. The program analyzes the training data and eliminates unneeded information to create a broad rule applied to future instances. For example, if the software is instructed to concentrate on the queen in chess, it will discard all non-immediate-effect pieces.

· The 90s – Emergence of various machine learning applications 

Scientists began to apply machine learning in data mining, adaptive software, web applications, text learning, and language learning in the 1990s. Scientists create computer programs that can analyze massive amounts of data and draw conclusions or learn from the findings. The term “Machine Learning” was coined as scientists were finally able to develop software in such a way that it could learn and improve on its own, requiring no human input.

· The Millennium – The rise of adaptive programming

The new millennium saw an unprecedented boom in adaptive programming. Machine learning went hand to hand with adaptive solutions for a long time. These programs can identify patterns, learn from experience, and improve themselves based on the feedback they receive from the environment.

Deep learning is an example of adaptive programming, where algorithms can “see” and distinguish objects in pictures and videos, which was the underlying technology behind Amazon GO shops. Customers are charged as they walk out without having to stand in line.

amazon-go
The history of Machine Learning: Amazon GO shops charge customers as they walk out without standing in line (Image courtesy of Amazon)

· Today – Machine learning is a valuable tool for all industries

Machine learning is one of today’s cutting-edge technologies that has aided us in improving not just industrial and professional procedures but also day-to-day life. This branch of machine learning uses statistical methods to create intelligent computer systems capable of learning from data sources accessible to it.

Contrary to popular belief, the history of machine learning, which enables machines to learn tasks for which they are not specifically programmed
The history of machine learning: Medical diagnosis is one area that ML will change soon

Machine learning is already being utilized in various areas and sectors. Medical diagnosis, image processing, prediction, classification, learning association, and regression are just a few applications. Machine learning algorithms are capable of learning from previous experiences or historical data. Machine learning programs use the experience to produce outcomes.

Organizations use machine learning to gain insight into consumer trends and operational patterns, as well as the creation of new products. Many of today’s top businesses incorporate machine learning into their daily operations. For many businesses, machine learning has become a significant competitive differentiator. In fact, machine learning engineering is a rising area.

· Tomorrow – The future of Machine Learning: Chasing the quantum advantage

Actually, our article was supposed to end here, since we came to today in the history of machine learning, but it doesn’t, because tomorrow holds more…

For example, Quantum Machine Learning (QML) is a young theoretical field investigating the interaction between quantum computing and machine learning methods. Quantum computing has recently been shown to have advantages for machine learning in several experiments. The overall objective of Quantum Machine Learning is to make things move faster by combining what we know about quantum computing with conventional machine learning. The idea of Quantum Machine Learning is derived from classical Machine Learning theory and interpreted in that light.

The application of quantum computers in the real world has advanced rapidly during the last decade, with the potential benefit becoming more apparent. One important area of research is how quantum computers may affect machine learning. It’s recently been demonstrated experimentally that quantum computers can solve problems with complex correlations between inputs that are difficult for traditional systems.

According to Google’s research, quantum computers may be more beneficial in certain applications. Quantum models generated on quantum computing machines might be far more potent for particular tasks, allowing for quicker processing and generalization on fewer data. As a result, it’s crucial to figure out when such a quantum edge can be exploited…

]]>
https://dataconomy.ru/2022/04/27/the-history-of-machine-learning/feed/ 0
The History of Data Mining https://dataconomy.ru/2016/06/16/history-data-mining/ https://dataconomy.ru/2016/06/16/history-data-mining/#comments Thu, 16 Jun 2016 08:00:39 +0000 https://dataconomy.ru/?p=15909 Data mining is everywhere, but its story starts many years before Moneyball and Edward Snowden. The following are major milestones and “firsts” in the history of data mining plus how it’s evolved and blended with data science and big data. Data mining is the computational process of exploring and uncovering patterns in large data sets a.k.a. […]]]>

Data mining is everywhere, but its story starts many years before Moneyball and Edward Snowden. The following are major milestones and “firsts” in the history of data mining plus how it’s evolved and blended with data science and big data.

Data mining is the computational process of exploring and uncovering patterns in large data sets a.k.a. Big Data. It’s a subfield of computer science which blends many techniques from statistics, data science, database theory and machine learning.

data_mining_timeline

 

 

1763 Thomas Bayes’ paper is published posthumously regarding a theorem for relating current probability to prior probability called the Bayes’ theorem. It is fundamental to data mining and probability, since it allows understanding of complex realities based on estimated probabilities.

1805 Adrien-Marie Legendre and Carl Friedrich Gauss apply regression to determine the orbits of bodies about the Sun (comets and planets). The goal of regression analysis is to estimate the relationships among variables, and the specific method they used in this case is the method of least squares. Regression is one of key tools in data mining.

1936 This is the dawn of computer age which makes possible the collection and processing of large amounts of data. In a 1936 paper, On Computable Numbers, Alan Turing introduced the idea of a Universal Machine capable of performing computations like our modern day computers. The modern day computer is built on the concepts pioneered by Turing.

1943 Warren McCulloch and Walter Pitts were the first to create a conceptual model of a neural network. In a paper entitled A logical calculus of the ideas immanent in nervous activity, they describe the idea of a neuron in a network. Each of these neurons can do 3 things: receive inputs, process inputs and generate output.

1965 Lawrence J. Fogel formed a new company called Decision Science, Inc. for applications of evolutionary programming. It was the first company specifically applying evolutionary computation to solve real-world problems.

1970s With sophisticated database management systems, it’s possible to store and query terabytes and petabytes of data. In addition, data warehouses allow users to move from a transaction-oriented way of thinking to a more analytical way of viewing the data. However, extracting sophisticated insights from these data warehouses of multidimensional models is very limited.

1975 John Henry Holland wrote Adaptation in Natural and Artificial Systems, the ground-breaking book on genetic algorithms. It is the book that initiated this field of study, presenting the theoretical foundations and exploring applications.

1980s HNC trademarks the phrase “database mining.” The trademark was meant to protect a product called DataBase Mining Workstation. It was a general purpose tool for building neural network models and now no longer is available. It’s also during this period that sophisticated algorithms can “learn” relationships from data that allow subject matter experts to reason about what the relationships mean.

1989 The term “Knowledge Discovery in Databases” (KDD) is coined by Gregory Piatetsky-Shapiro. It also at this time that he co-founds the first workshop also named KDD.

1990s The term “data mining” appeared in the database community. Retail companies and the financial community are using data mining to analyze data and recognize trends to increase their customer base, predict fluctuations in interest rates, stock prices, customer demand.

1992 Bernhard E. Boser, Isabelle M. Guyon and Vladimir N. Vapnik suggested an improvement on the original support vector machine which allows for the creation of nonlinear classifiers. Support vector machines are a supervised learning approach that analyzes data and recognizes patterns used for classification and regression analysis.

1993 Gregory Piatetsky-Shapiro starts the newsletter Knowledge Discovery Nuggets (KDnuggets). It was originally meant to connect researchers who attended the KDD workshop. However, KDnuggets.com seems to have a much wider audience now.

2001 Although the term data science has existed since 1960s, it wasn’t until 2001 that William S. Cleveland introduced it as an independent discipline. As per Build Data Science Teams, DJ Patil and Jeff Hammerbacher then used the term to describe their roles at LinkedIn and Facebook.

2003 Moneyball, by Michael Lewis, is published and changed the way many major league front offices do business. The Oakland Athletics used a statistical, data-driven approach to select for qualities in players that were undervalued and cheaper to obtain. In this manner, they successfully assembled a team that brought them to the 2002 and 2003 playoffs with 1/3 the payroll.

Present In February 2015, DJ Patil became the first Chief Data Scientist at the White House. Today, data mining is widespread in business, science, engineering and medicine just to name a few. Mining of credit card transactions, stock market movements, national security, genome sequencing and clinical trials are just the tip of the iceberg for data mining applications. Terms like Big Data are now commonplace with the collection of data becoming cheaper and the proliferation of devices capable of collecting data.

This content originally appeared on http://rayli.net/blog/

Like this article? Subscribe to our weekly newsletter to never miss out!

]]>
https://dataconomy.ru/2016/06/16/history-data-mining/feed/ 11
Beginner’s Guide to the History of Data Science https://dataconomy.ru/2016/03/11/beginners-guide-history-data-science/ https://dataconomy.ru/2016/03/11/beginners-guide-history-data-science/#comments Fri, 11 Mar 2016 08:30:02 +0000 https://dataconomy.ru/?p=15092 “Big data” and “data science” may be some of the bigger buzzwords this decade, but they aren’t necessarily new concepts. The idea of data science spans many different fields, and has been slowly making its way into the mainstream for over fifty years. In fact, many considered last year the fiftieth anniversary of its official […]]]>

“Big data” and “data science” may be some of the bigger buzzwords this decade, but they aren’t necessarily new concepts. The idea of data science spans many different fields, and has been slowly making its way into the mainstream for over fifty years. In fact, many considered last year the fiftieth anniversary of its official introduction. While many proponents have taken up the stick, made new assertions and challenges, there are a few names and dates you need know.

1962. John Tukey writes “The Future of Data Analysis.” Published in The Annals of Mathematical Statistics, a major venue for statistical research, he brought the relationship between statistics and analysis into question. One famous quote has since struck a chord with modern data lovers:

“For a long time I have thought I was a statistician, interested in inferences from the particular to the general. But as I have watched mathematical statistics evolve, I have had cause to wonder and to doubt…I have come to feel that my central interest is in data analysis, which I take to include, among other things: procedures for analyzing data, techniques for interpreting the results of such procedures, ways of planning the gathering of data to make its analysis easier, more precise or more accurate, and all the machinery and results of (mathematical) statistics which apply to analyzing data.”

1974. After Tukey, there is another important name that any data enthusiast should know: Peter Naur. He published the Concise Survey of Computer Methods, which surveyed data processing methods across a wide variety of applications. More importantly, the very term “data science” is used repeatedly. Naur offers his own definition of the term: “The science of dealing with data, once they have been established, while the relation of the data to what they represent is delegated to other fields and sciences.” It would take some time for the ideas to really catch on, but the general push toward data science started to pop up more and more often after his paper.

1977. The International Association for Statistical Computing (IASC) was founded. Their mission was to “link traditional statistical methodology, modern computer technology, and the knowledge of domain experts in order to convert data into information and knowledge.” In this year, Tukey also published a second major work: “Exploratory Data Analysis.” Here, he argues that emphasis should be placed on using data to suggest hypotheses for testing, and that exploratory data analysis should work side-by-side with confirmatory data analysis. In 1989, the first Knowledge Discovery in Databases (KDD) workshop was organized, which would become the annual ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD).

In 1994 the early forms of modern marketing began to appear. One example comes from the Business Week cover story “Database Marketing.” Here, readers get the news that companies are gathering all kinds of data in order to start new marketing campaigns. While companies had yet to figure out what to do with all of the data, the ominous line that “still, many companies believe they have no choice but to brave the database-marketing frontier” marked the beginning of an era.

In 1996, the term “data science” appeared for the first time at the International Federation of Classification Societies in Japan. The topic? “Data science, classification, and related methods.” The next year, in 1997, C.F. Jeff Wu gave an inaugural lecture titled simply “Statistics = Data Science?

Already in 1999, we get a glimpse of the burgeoning field of big data. Jacob Zahavi, quoted in “Mining Data for Nuggets of Knowledge” in Knowledge@Wharton had some more insight that would only prove to true over the following years:

“Conventional statistical methods work well with small data sets. Today’s databases, however, can involve millions of rows and scores of columns of data… Scalability is a huge issue in data mining. Another technical challenge is developing models that can do a better job analyzing data, detecting non-linear relationships and interaction between elements… Special data mining tools may have to be developed to address web-site decisions.”

And this was only in 1999! 2001 brought even more, including the first usage of “software as a service,” the fundamental concept behind cloud-based applications. Data science and big data seemed to grow and work perfectly with the developing technology. One of the many more important names is William S. Cleveland. He co-edited Tukey’s collected works, developed valuable statistical methods, and published the paper “Data Science: An Action Plan for Expanding the Technical Areas of the field of Statistics.” Cleveland put forward the notion that data science was an independent discipline and named six areas in which he believed data scientists should be educated: multidisciplinary investigations, models and methods for data, computing with data, pedagogy, tool evaluation, and theory.

2008. The term “data scientist” is often attributed to Jeff Hammerbacher and DJ Patil, of Facebook and LinkedIn—because they carefully chose it. Attempting to describe their teams and work, they settled on “data scientist” and a buzzword was born. (Oh, and Patil continues to make waves as the current Chief Data Scientist at White House Office of Science and Technology Policy).

2010. The term “data science” has fully infiltrated the vernacular. Between just 2011 and 2012, “data scientist” job listings increased 15,000%. There has also been an increase in conferences and meetups devoted solely to data science and big data. The theme of data science hasn’t only become popular by this point, it has become highly developed and incredibly useful.

2013 was the year data got really big. IBM shared statistics that showed 90% of the world’s data had been created in the preceding two years, alone.

2016 may have only just began, but predictions are already begin made for the upcoming year. Data science is entrenched in machine learning, and many expect this to be the year of Deep Learning. With access to vast amounts of data, deep learning will be key towards moving forward into new areas. This will go hand-in-hand with opening up data and creating open source data solutions that enable non-experts to take part in the data science revolution.

In the past decade, the idea of data science exploded and slowly became what we recognize today. One vital point analysts understand is that data science and big data are not simply “scaling up” data. Instead, it means a shift in study and analysis. Despite seeming almost completely ordinary in today’s world, like something that could not possibly be removed from research and study, the nature and importance of data science was not always so clear, and its exact nature will continue to develop alongside technology.

image credit: Jer Thorp

Like this article? Subscribe to our weekly newsletter to never miss out!

]]>
https://dataconomy.ru/2016/03/11/beginners-guide-history-data-science/feed/ 17