MLG 002 Difference Between Artificial Intelligence, Machine Learning, Data Science

Feb 09, 2017 (updated Nov 23, 2021)
Click to Play Episode

Artificial intelligence is the automation of tasks that require human intelligence, encompassing fields like natural language processing, perception, planning, and robotics, with machine learning emerging as the primary method to recognize patterns in data and make predictions. Data science serves as the overarching discipline that includes artificial intelligence and machine learning, focusing broadly on extracting knowledge and actionable insights from data using scientific and computational methods.

Resources
Resources best viewed here
Andrew Ng - Machine Learning Specialization
An Introduction to Statistical Learning (ISLR) (2nd Edition)
Show Notes
CTA

Sitting for hours drains energy and focus. A walking desk boosts alertness, helping you retain complex ML topics more effectively.Boost focus and energy to learn faster and retain more.Discover the benefitsDiscover the benefits

Data Science Overview

  • Data science encompasses any professional role that deals extensively with data, including but not limited to artificial intelligence and machine learning.
  • The data science pipeline includes data ingestion, storage, cleaning (feature engineering), and outputs in data analytics, business intelligence, or machine learning.
  • A data lake aggregates raw data from multiple sources, while a feature store holds cleaned and transformed data, prepared for analysis or model training.
  • Data analysts and business intelligence professionals work primarily with data warehouses to generate human-readable reports, while machine learning engineers use transformed data to build and deploy predictive models.
  • At smaller organizations, one person ("data scientist") may perform all data pipeline roles, whereas at large organizations, each phase may be specialized.
  • Wikipedia: Data Science describes data science as the interdisciplinary field for extracting knowledge and insights from structured and unstructured data.

Artificial Intelligence: Definition and Sub-disciplines

  • Artificial intelligence (AI) refers to the theory and development of computer systems capable of performing tasks that typically require human intelligence, such as visual perception, speech recognition, decision-making, and language translation. (Wikipedia: Artificial Intelligence)
  • The AI discipline is divided into subfields:
    • Reasoning and problem solving
    • Knowledge representation (such as using ontologies or knowledge graphs)
    • Planning (selecting actions in an environment, e.g., chess- or Go-playing bots, self-driving cars)
    • Learning
    • Natural language processing (simulated language, machine translation, chatbots, speech recognition, question answering, summarization)
    • Perception (AI perceives the world with sensors; e.g., cameras, microphones in self-driving cars)
    • Motion and manipulation (robotics, transforming decisions into physical actions via actuators)
    • Social intelligence (AI tuned to human emotions, sentiment analysis, emotion recognition)
    • General intelligence (Artificial General Intelligence, or AGI: a system that generalizes across all domains at or beyond human skill)
  • Applications of AI include autonomous vehicles, medical diagnosis, creating art, proving theorems, playing strategy games, search engines, digital assistants, image recognition, spam filtering, judicial decision prediction, and targeted online advertising.
  • AI has both objective definitions (automation of intellectual tasks) and subjective debates around the threshold for "intelligence."
  • The Turing Test posits that if a human cannot distinguish an AI from another human through conversation, the AI can be considered intelligent.
  • Weak AI targets specific domains, while general AI aspires to domain-independent capability.
  • AlphaGo Movie depicts the use of AI planning and learning in the game of Go.

Machine Learning: Within AI

  • Machine learning (ML) is a subdiscipline of AI focused on building models that learn patterns from data and make predictions or decisions. (Wikipedia: Machine Learning)
  • Machine learning involves feeding data (such as spreadsheets of stock prices) into algorithms that detect patterns (learning phase) and generate models, which are then used to predict future outcomes.
  • Although ML started as a distinct subfield, in recent years it has subsumed many of the original AI subdisciplines, becoming the primary approach in areas like natural language processing, computer vision, reasoning, and planning.
  • Deep learning has driven this shift, employing techniques such as neural networks, convolutional networks (image processing), and transformers (language tasks), allowing generalizable solutions across multiple domains.
  • Reinforcement learning, a form of machine learning, enables AI systems to learn sequences of actions in complex environments, such as games or real-world robotics, by maximizing cumulative rewards.
  • Modern unified ML models, such as Google’s Pathways and transformer architectures, can now tackle tasks in multiple subdomains (vision, language, decision-making) with a single framework.

Data Pipeline and Roles in Data Science

  • Data engineering covers obtaining and storing raw data from various data sources (datasets, databases, streams), aggregating into data lakes, and applying schema or permissions.
  • Feature engineering cleans and transforms raw data (imputation, feature transformation, selection) for machine learning or analytics.
  • Data warehouses store column-oriented, recent slices of data optimized for fast querying and are used by analysts and business intelligence professionals.
  • The analytics branch (data analysts, BI professionals) uses cleaned, curated data to generate human insights and reports.
    • Data analysts apply technical and coding skills, while BI professionals often use specialized tools (e.g., Tableau, Power BI).
  • The machine learning branch uses feature data to train predictive models, automate decisions, and in some cases, trigger actions (robots, recommender systems).
  • The role of a "data scientist" can range from specialist to generalist, depending on team size and industry focus.

Historical Context of Artificial Intelligence

  • Early concepts of artificial intelligence appear in Greek mythology (automatons) and Jewish mythology (Golems).
  • Ramon Lull in the 13th century and Leonardo da Vinci constructed early automatons.
  • Contributions:
    • Thomas Bayes (probability inference, 1700s)
    • George Boole (logical reasoning, binary algebra)
    • Gottlob Frege (propositional logic)
    • Charles Babbage and Ada Byron/Lovelace (Analytical Engine, 1832)
    • Alan Turing (Universal Turing Machine, 1936; foundational ideas on computing and AI)
    • John von Neumann (Universal Computing Machine, 1946)
    • Warren McCulloch, Walter Pitts, Frank Rosenblatt (artificial neurons, perceptron, foundation of connectionist/neural net models)
    • John McCarthy, Marvin Minsky, Arthur Samuel, Oliver Selfridge, Ray Solomonoff, Allen Newell, Herbert Simon (Dartmouth Workshop, 1956: "AI" coined)
    • Newell and Simon (Heuristics, General Problem Solver)
    • Feigenbaum (expert systems)
    • GOFAI/symbolism (logic- and knowledge-based systems)
  • The “AI winter” followed the Lighthill report (1970s) due to overpromising and slow real-world progress.
  • AI resurgence in the 1990s was fueled by advances in computation, increased availability of data (the era of "big data"), and improvements in neural network methodologies (notably Geoffrey Hinton's optimization of backpropagation in 2006).
  • The 2010s saw dramatic progress, with companies such as DeepMind (acquired by Google in 2014) achieving state-of-the-art results in reinforcement learning and general AI research.
  • The Sub-disciplines of AI and other resources:

Further Learning Resources

CTA

Go from concept to action plan. Get expert, confidential guidance on your specific AI implementation challenges in a private, one-hour strategy session with Tyler.Get personalized guidance from Tyler to solve your company's AI implementation challenges.Book Your Session with TylerBook Your Call with Tyler

Transcript
Machine learning guide episode two. This is what is artificial intelligence, machine learning and data science. Let's start with a bird's eye view. Data science is the all encompassing umbrella term, inside of which is artificial intelligence, inside of which is machine learning. So data science contains ai, and AI contains ml. But before we go into data science, let's actually start with artificial intelligence. From Oxford Dictionary, AI means the theory and development of computer systems able to perform tasks that normally require human intelligence, such as visual perception, speech recognition, decision making, and translation between languages. So basically AI is automated intelligence automation of things that normally require human intelligence. That's the objective definition of ai. In the same way that the Industrial Revolution was automating the body through robotics and machinery, the AI revolution is automating the brain by way of intellect. So the objective definition is pretty clear, but in a sense it may not be that helpful. And so people have difficulty defining AI because they have in mind more of a subjective take on it. And that subjective definition of artificial intelligence can be difficult because defining intelligence is difficult. So we have the objective definition of ai. We have the subjective definition of ai. And since neither of those takes are super helpful in this context, let's instead just break AI down into its sub-fields so that you have a better understanding of AI as a discipline, AI is broken down into multiple sub-disciplines, straight from Wikipedia. These are reasoning and problem solving knowledge, representation, planning, learning, natural language processing, perception, motion and manipulation, social intelligence and general intelligence. Let's start with natural language processing or NLP. And the reason I'm gonna start there is I'm gonna kind of paint a storyline of the disciplines of ai, and this is a good starting point. NLP is the sub-discipline of AI dedicated to language. Anything that is simulated language is natural language processing. This can include machine translation, for example, Google Translate from English to Spanish. It can include chat bots, things like Siri for example. Or let's say that you have a customer service chat bot on a website and you're chatting with that bot. You can either be speaking your utterances out loud and it will perform speech to text that's within the domain of NLP. And then once it has that text, it's going to perform something called named entity recognition. So if a chat bot says, how can I help you today? And you say, I need help with my credit card, it'll pull out credit card. I. As a named entity that's named Entity Recognition or NER. We'll talk about these kinds of topics in NLP in a future episode, and it can do question answering, summarization and any other task associated with language. Next, let's talk about the subdiscipline knowledge representation. A long time ago, IBM created a jeopardy playing bot called Deep Blue. What would eventually become IBM Watson and Deep Blue had inside of it the capacity to play jeopardy, not just by way of natural language processing. You need NLP so that when the question is asked, it can take that question and transform it into something that can be used internally to answer the question. But also that there has to be some internal knowledge representation, either trained on the web or on Wikipedia or on books, some way of taking this knowledge about the world out there and transforming it into an internal representation. Usually in this sub-discipline, we deal with things called ontologies or knowledge graphs, so that after you've gotten the question and transformed it into the type of representation, you can traverse the knowledge graph in order to find the answer, come back with the answer, and then represent that answer using NLP back to the human. So these two fields are somewhat related, especially these days. In NLP, we have a specific task called question answering, where you can hand the agent a corpus of documents, for example, a book, and you can ask it a question and it will answer the question either attractively by pulling the most probabilistic answer out of that corpus or abstractly, such that it'll actually parse your question, figure out what you implied by it. Go through all the documents in the corpus in order to paraphrase an answer to the best of its quote unquote understanding, and then relay that answer back to you. That's called abstractive question answering, which to me is a little bit more magical than extractive question answering. These sound a lot like knowledge representation, but they're slightly different domains. Again, knowledge representation isn't explicitly dealing in language per se, although it will certainly be using language as part of its tool chain. Instead, it's the sub-discipline of representing knowledge in the AI agent such that that knowledge can be traversed for the correct answer rather than just by pre-trained text extraction language models. Next up, we have reasoning and problem solving. This is related to the prior sub-discipline of knowledge, representation, reasoning and problem solving though are more typically dealing with probabilistic models about the world in order to solve problems. Then we have planning. Planning is actually the subdiscipline of taking actions in an environment by an artificial agent. This is what you may have heard of with chess and go playing bots. So a long time ago, AI researchers and developers created a bot that could play chess and played chess against the current reigning champion at the time, Gary Kasparov and beat him. And so that created a lot of publicity around ai. Recently, Google's DeepMind Alpha Go learns to play the Game of Go and was pitted against the reigning champion, Lisa Dole. And one, again, bringing a lot of publicity to AI by way of much more magical tooling that I'll talk about in a bit than was previously used in the case of chess. It was a great movie about DeepMind versus Lisa Doll, which I'll link in the show notes. I highly recommend watching it. So planning is more about taking actions in an environment, whether that be playing video games or driving a self-driving car where the car needs to be able to take turn by turn actions on the roads. Speaking of self-driving cars, taking actions on the roads, there's two other pieces that need to be present for a car. The first is perception. So an AI needs to be able to perceive the world around it, eyes, ears, mouth, and nose and touch. So a self-driving car needs to be able to see the road by way of cameras or lidar. It may also need to be able to hear the road hear horns blaring or cars screeching. And so that would be using microphones. So this is perception. It also needs to be able to perform actions inside of its environment. This is the sub-discipline called motion and manipulation, which is basically just robotics. So a car has seen the road through a camera. That pixel data gets transformed by AI into some internal representation. It performs some planning steps in order to determine whether it should break, accelerate, turn left or turn right, and then it has to be able to transform those action steps into physical reality. By way of robotics, which is the subdiscipline motion and manipulation, and in particular we call this hardware, the point at which the AI performs physical actions in an environment, actuators. So an actuator in a self-driving car might be the brakes, the steering wheel, the wheels, whatever. Now, you'll notice I'm talking about robotics here, and a large part of robotics is simply hardware. That's cameras, microphones, and the machinery used in robotics. But we do include robotics as a subdiscipline of AI to the extent that. Interfacing with the hardware by way of intelligence, whether that be planning in the case of taking actions or representing what it perceives in a way that can be internally utilized by the AI agent in the case of perception. Now you've probably seen in the news all the cool stuff by Boston Dynamics. Those guys are killing it in the robotics field. We have robots that are jumping on boxes and doing backflips. There's one thing they can't do. Well defend themselves against hockey sticks. I always joke that if there ever was an evil killer robot takeover, well we know their weakness because Boston Dynamics has shown us that's hockey sticks. So you might wanna stock up on hockey sticks to defend yourself against the AI revolution. Also listed on Wikipedia is social intelligence and general intelligence. Social intelligence is the concept of AI being tuned into human emotions. So we may be trying to follow a human's emotional flow through time. In the case of robo therapists, I myself am actually building a journal app that uses AI to give you resources and recommendations through your journaling journey. You'll hear about it in future episodes. It's called No the, and there are apps which are robo therapists that you can converse with, and these will attempts to pick up on your emotional state based on NLP tasks, such as sentiment analysis, whether the things you're typing are positive or negative or neutral. And there are also computer vision technologies, which we'll try to pick up on the emotions on a human's face. And then finally, general intelligence or artificial general intelligence, A-G-I-A-G-I is the pot of gold at the end of the rainbow. It's a pipe dream. It's a pie in the sky. It's the goal of creating an artificial intelligence that is not only good at its particular tasks, but is also generalizable to all other domains so that it is as effective at those as a human, if not better. Okay, so we say weak ai. If the AI is dedicated to a specific domain, for example, Siri is really relegated to natural language processing. A Roomba, it's mostly robotics, but it's also incorporating action and planning. That's a twofer right there. Self-driving cars. We're getting a little bit closer to the whole shebang, the grandeur prize of combining all of these AI subdisciplines into one. We have perception, namely computer vision and hearing to some extent. On the one hand, it may be listening to sounds on the road. It can also listen to voice commands inside the car. Of course. Action and planning. That's turn by turn, navigation, acceleration, braking, and then motion and manipulation all through robotics. So a Tesla self-driving car is getting us closer to the end goal of artificial general intelligence, but as you can tell, it is not very general. The word general in this context means that the robot or the AI agent should be able to generally apply itself to any context in where in which a human can find themself and perform or act as intelligently as a human, if not better. So let's say we take a Boston Dynamics robot, which is good at jumping on boxes and doing back flips and hauling luggage like a pack meal, and then put it in front of a court so that it's handling legal cases. It should be able to pick up the skills needed to handle legal cases as well, or better than a human who is plopped down into the court. And that's the generality of artificial general intelligence. Weak means it's dedicated to a specific subdomain. Not necessarily that it's weak at its task, but just that it can only handle a specific domain. And general means that it can handle all domains and generalize across contexts universally like a human could. Now, when I say pie in the sky pipe dream, I don't mean that this isn't gonna happen. I actually do believe we will achieve a GI, and I believe that it will happen in my lifetime between Google Mind, open ai, and all the other FANG companies go and hog wild. With AI these days, it looks like we're on a pretty focused path with major improvements week by week. What I mean by that is that a GI is more of a concept and a future goal, rather than the day-to-day dealings of somebody who works in artificial intelligence and machine learning. Basically the objective definition of AI is automated intelligence or automating intellectual tasks, which typically require human intelligence to perform. That's the objective definition, but it's not very helpful because, for example, I could write a Python script that does some number crunching on a spreadsheet that is automation of an intellectual task that usually requires human intelligence. And so that segues us into the subjective definition of artificial intelligence, which is really hard to pin down. You'll find that when you ask somebody what is artificial intelligence, whether you can read the definition from the dictionary, but strangely it can be easy to get into something of a heated debate of about what the actual definition of AI is. You'll find that a lot of people take the definition of AI and make it a little personal and make it subjective. And the reason for this, I think, is because what they're really driving after, rather than the definition of AI, is the definition of I, the definition of intelligence in general. Indeed defining intelligence can be quite a subjective and dicey matter, which segues into the conversation of what is consciousness. For example, I discuss intelligence and consciousness in a future episode, but let's do a little bit of hand waving here so that we can become a little bit more settled with the subjective take on artificial intelligence. The way I see it, we have manual hard-coded scripts On the one hand, on the far left. And then we have true intelligence all the way to the right far on the horizon past what the eye can see. And the reason I say it this way is because I think that intelligence is an analog concept in the same way that some humans can be more intelligent than others, and humans are more intelligent than dogs and dogs than cats and cats than rats and so on. There's sort of an analog scale to intelligence. And so too will be the case with artificial intelligence. And what is the upper limit? We don't know. And therefore there is no sort of end to the, to the right side of the scale, it's analog. So on the left of the scale, we have handwritten scripts that can be considered AI by the objective definition. And on the far right we have the pie in the sky concept of a GI, which I'll discuss in a bit. And anywhere in the middle is sort of what we're dealing with in trying to define AI subjectively. The layers by which the software becomes removed from the programmer adds elements of magic to the concept of ai. So a handwritten script is not very magical. A machine learning model, which we'll discuss later, that learns, let's say, the rules of the game of chess or go, or how to drive a self-driving car down the road. Well, that sum number of layers removed from the original programmer. And that removal adds an element of magic. That magic is the sense by which the concept of a AI can be defined subjectively. So where do we draw the line by which we call something magically AI in the sense that we're all looking for, well, Alan Turing had a really good take here. The Turing Test or the Imitation Game. Alan Turing is essentially the creator, conceptually of the computer by way of the Universal touring machine. He was fascinated by the theoretical concept of artificial intelligence. Of course, technically achieving AI was well outside of the technological capabilities of his time, but he was really interested in the concept and he came up with the Turing test, which basically says if you are chatting with an artificial agent, an ai, basically he envisioned that you're on a terminal with a chatbot and the other end of the chatbot responding to your utterances you don't know is either a human or an artificially intelligent agent. If you're chatting with this agent and you can't tell if it's a human or an AI than it is intelligent. Whether it's the human or the ai, if an AI can sufficiently convince you that it is intelligent through a chat bot where you're poking and prodding and trying to find the holes and you can't find them, then it is intelligent in the magical sense. So he says, if an AI can convince you that it's not, not an AI than it is, than it is intelligent. So in a way, if we say that there's a sliding scale from hard scripting on the left and artificial general intelligence way into the horizon to infinity, the point at which we achieve intelligence in the magical sense that we're all trying to subjectively pin down is the point at which you can't deny it. And it's as simple as that. If it walks like a duck and it talks like a duck, which is behaviorism, if you experience an AI agent and you are flabbergasted. And can't deny that it is intelligence than it is, and anything up until that point is not, and anything beyond that point is the future. So that's sort of the marker delineating this analog infinity scale as the point at which we subjectively define intelligence. That is, you'll know it when you see it. So that finally takes us to learning machine learning. The last subdiscipline listed for artificial intelligence and the most interesting and powerful subdiscipline of artificial intelligence, as you'll see here, the reason I saved it for last is not only that, it's the primary subject of this podcast series, but also because it is machine learning, this subdiscipline of AI that makes disambiguating the terms AI and machine learning difficult. And that is because in recent times, machine learning has been subsuming, all of the other subdisciplines of artificial intelligence, machine learning what was previously relegated to its own siloed subdiscipline. Is now breaking out of its cage and hostile taking over all the other subdisciplines, grabbing a choke hold on. Those subdisciplines, and some of these subfields have been completely turned blue. There's nothing left, but machine learning. So let's define machine learning first and then let's revisit the subfields. Machine learning is learning on data to make predictions. It's pattern recognition over data so that future predictions can be made. It's learning on data. That's what machine learning is. It's really simple. It's really easy to understand. The idea is you feed it a spreadsheet. Let's say that spreadsheet is stock prices for a particular stock over time, and what we're trying to do is predict tomorrow's price for that stock. Okay? It's a typical use case of machine learning, algorithmic trading bots. You take this spreadsheet, this data. The machine learning algorithm will look at all the data in the spreadsheet in order to try to find patterns in the data, is trying to learn the patterns of the data. What is it that makes the price go up or go down given other variables in that spreadsheet, such as the time of day, day of week, sentiment on Twitter, whatever the case, learn those patterns. And once we have those learned patterns saved, we combine the learned patterns into the algorithm into what's called a model. A machine learning model. So a model is the combination of the algorithm, what the computer programmer wrote, and the patterns learned, what the machine learning model itself picked up. Now we have a model and that model can make predictions for the future. So the model can now predict based on today's context or circumstances, what's the price gonna be for this stock tomorrow. And then it may either just predict the price or it may actually take action on that prediction such as to buy or sell on that stock. So that's machine learning. It is learning on data to make predictions. It's as simple as that. Now, previously, like I said, this was a dedicated subdiscipline of AI that was focused primarily maybe on tabular data, learning on spreadsheets or databases, the patterns in that tabular data by way of simple algorithms, simple models, things like linear and logistic regression, which we'll discuss later. And these simple models made machine learning maybe less interesting to the other practitioners in the other sub-disciplines of ai. But after the deep learning revolution of about 2015, machine learning started to become generally applicable and started dominating the other sub-disciplines of artificial intelligence. So natural language processing. Previously NLP was language theory. Basically, you go to university, you study language theory. You learn all about syntax trees and the ways in which certain words combine with other words in grammatical constructs so that you could traverse these trees in order to pull out named entities or answers to questions. These were hard-coded scripts by domain experts in NLP. Well, along comes machine learning with its recurrent neural networks and transformers models and clobbers. These original. Expert crafted scripts where each expert was a sub expert within NLP. Maybe some expert and their script were really good at question answering. And another for named entity recognition and another for machine translation. Well, these Transformers models are generally applicable, generally applicable to all of the tasks within NLP. Transformers can handle machine translation, question answering, summarization and more, all with a universal learning model. And that this model can be trained on the language tasks generally, rather than handcrafted on a specific language task. For example, the difference between legal documents versus news articles. So machine learning takes us some steps away from a handcrafted script, which adds some magic to our subdiscipline of ai and some generality a step towards artificial general intelligence. Some generality insofar as a transformers model can be generally applied to various tasks and domains within NLP. Computer vision. Previously computer vision was handcrafted scripts, algorithms that had hand designed pattern matching patches. These sliding windows that would slide over an image from top to bottom, left to right, where the patterns that it's looking for in order to detect an object or to classify an image were domain. Expert handcrafted pattern matching windows. So an expert would design a series of manual convolutions, which were looking for lines and edges, or looking for certain patterns for dog breeds. For example, some such algorithms were called hog, HOG, and har, HAAR, which I'll discuss in the computer vision episode. Well along comes machine learning with its convolutional neural networks. These CNNs learn the convolutions, these sliding windows, these pattern identifying patches and blows. The old guard of handcrafted solutions to arenas. Again, these models are not only layers removed from handcrafted solutions, insofar as they can learn the pixel pattern matching process, but also are generalizable across different domains. So the hog and horror people might have had to handcraft different convolutions for different domains, whether we're dealing with animal matchers or clothing matchers or outdoors objects, matchers, any one convolutional neural network can handle all these contexts generally. Then finally, reasoning, problem solving, planning, motion, and manipulation and perception. These were all previously dedicated subdisciplines in AI and they still are to some extent. In particular, the reasoning. Subdiscipline dealt a lot in probabilistic models and the planning subdiscipline dealt a lot in trees. Search trees like the A star search tree and these types of trees would've been used, for example, in the chess playing bot against Gary Kasparov. Well along comes deep reinforcement learning models with their Deep Q Networks and their proximate policy optimization models. And I don't know which specific reinforcement models are used in different contexts. For example, what reinforcement model, what reinforcement learning models used in AlphaGo Zero. But it is becoming very obvious that there is a massive amount of generalizability for use of deep reinforcement learning models in any situation that requires taking steps in an environment, planning actions, and taking those actions. We previously had search trees and markoff models and all these things which would've been custom tailored to specific environments. For example, pre-programmed with the rules of chess, along with the how-tos of playing chess. Well, deep reinforcement learning models not only learn how to play a game or drive on a road by taking actions in an environment and observing the point value consequences downstream at the end of the game in order to modify its behavior, but it also learns how the game is played in general. What game is even being played? So many of these reinforcement learning models, all you have to do is show it a screen through time. So it takes space in pixel values, and time, the progression of the video game as it takes specific actions, and it will learn the rules of the game, the context and environment of the game, and how to play it. So now at this point, we're another layer removed from a handcrafted script, even more magical then than the models used in NLP and computer vision. So machine learning is not only more effective at tackling a lot of these subdisciplines in AI than the old guard of previously handcrafted scripts, but it's also more generalizable and frankly more magical. And speaking of generalizability. Strides are being made on many fronts now to unify these machine learning models to tackle multiple subdomains within AI at once. For example, recently in the news, Google announced something called Pathways, which as I understand it is a machine learning model that can be applied to computer vision, natural language processing, and reinforcement learning. Also, as you'll learn in the NLP episodes, an NLP model concept called transformers is now also being used for vision and other tasks. So steps are being taken towards a grand unification machine learning model. And those steps also take us towards dominating artificial intelligence at large by way of machine learning. So that's what's special about machine learning is that it seems to be the master algorithm of artificial intelligence. And that's also why disambiguating, those two terms has been relatively difficult is because previously I. ML was simply a subdiscipline of ai, but in modern times, it's the dominating driving force of AI in all of its other subdisciplines. Now, let's talk about data science. Data science, like I said, is the umbrella term that encompasses artificial intelligence and machine learning. More than that, data science is the umbrella term that encompasses anything related to data. If you are a professional and your daily job is dealing with data, then you are a data scientist. This can include things like database administration, data analytics, business intelligence, machine learning, artificial intelligence, even something as simple as maybe a tax person. If you're dealing with spreadsheets day in day out, to some extent, you're a data scientist. So roles that primarily deal in data are roles within data science. But there are roles more understood to be data science proper. And that's what we'll discuss here by way of a data pipeline. Because the typical roles dealing in data science fall somewhere in this pipeline. A pipeline from left to right, beginning to end, the beginning of the data pipeline, gets the data from somewhere and pipes it into the machine. And the output of the pipeline is either data analytics, business intelligence, or machine learning. And then there's a whole bunch of machinery in the middle. So let's go down this pipeline journey. The first step of the pipeline is data ingestion, ingestion, which is to get your data from somewhere. This data can be in a spreadsheet like Microsoft Excel, Excel sx, or CSV files. It can be in a database like Postgres or my SQL. Those are relational database management systems. There are also non-relational database management systems, or no SQL, no SQL databases like MongoDB and Dynamo db. Or it can be from a stream of data on the internet. A stream means that the data is coming at you so much and so fast that you don't want to store it anywhere. You just wanna process it real time en route. You wanna receive your data, do stuff to it, and then probably throw it away. So an example of a stream of data would be Twitter. Twitter is what's called a stream or a fire hose of data. It's coming at you so fast 'cause there's so many tweets in the world that you're probably not gonna be storing this into a SQL database or a no SQL database and certainly not a spreadsheet. A data set typically means file formatted data like a spreadsheet. A data store typically means database formatted data like on Postgres or my SQL, and then we have a stream or fire hose. This comes into your pipeline and now you want to put all that data somewhere. You want to aggregate all of the data coming to some use case all together in what's called a data lake. A data lake is the aggregation of various data sets, data stores into one repository, one umbrella, and that data is gonna be dirty. It's gonna not be cleaned up yet. It's coming straight from the ingestion phase. And data lake technologies allow you to apply certain unifying aggregating. Features on top of your data. One of these might be permissions. So for example, if you want to collect spreadsheets and databases and tweets all into one unified repository called a data lake, then you can use data lake technology to apply permissioning on that data lake, such that various downstream data users or consumers can access this data lake all by way of a unified permissioning system. Another feature you might wanna apply to your data lake is a unifying schema. If much of your data has a lot of structure in common, even though they're coming from disparate data sources, you can create a data schema that you generally apply to the data sources in this data lake so that downstream consumers of the data can expect a certain formatted structure to the data. The next phase of the pipeline might be something called a feature store. Now, this can sometimes be wrapped up in the data lake phase, or they can be separate components of the pipeline. A feature store is the phase at which you take the data from the data lake and you clean it up and you store it in its cleaned up state. So you're gonna have data coming from Twitter, coming from databases, and it's gonna be dirty. There's gonna be missing entries, missing columns for certain rows. Some fields are in text format or date format, and as you'll find later, data analytics and machine learning don't like dates or strings. They want numbers. Machine learning and data analytics always wants numbers. And so this feature engineering phase, they call it feature engineering, is where we will perform some of these steps. One step might be to fill in missing values. We call this imputing. You'll learn about imputation strategies later. Another feature engineering step might be taking your date column and turning it into numbers. So for example, we can take a date column and turn that date string into time of day, day of week, week of year. And so now we transformed the date feature into three other features. We call this feature transformation. As well as simply deciding which features to keep and which ones aren't worth keeping. For example, in the case of Twitter, username might not be that valuable. We really are concerned with the date the tweets text, and maybe the engagement, like the number of replies or number of retweets. So we throw away the username. This is called feature selection, and this whole process is called feature engineering. And then we store all these engineered features as something of a checkpoint in our pipeline called a feature store. And feature stores also allow you to version your transformed features such that. If a machine learning engineer has been working with data in such and such a way based on the way that the data engineer had previously been feature engineering, and the data engineer makes a modification to that phase in the pipeline, well, the machine learning engineer won't be sideswiped because they will be using a specific version of the feature store. Okay? So data ingestion takes our data from data sets and data stores and streams and fire hoses. These are all called data sources, stores them all into a data lake, which is dirty, unclean, large data or big data. You've probably heard that term, big data. Anytime we're working with data that can't fit on a single server, it's big data. We may apply unified tooling on top of that lake, things like permissioning or a common data schema. And then we pull that data outta the lake. We clean up the data, a step called feature engineering. Those steps may be feature transformations that is altering one feature into a different feature or set of features, imputation that is filling in missing values, feature selection, that is deciding which features are important and which aren't. And then now we're done with that third phase and we're ready to move on in the pipeline. At this phase in the pipeline, we're gonna veer left and right. There's a fork in the road just like there is with pipes. And so to the left we have data analytics and business intelligence. And to the right we have machine learning. These are two typically separate roles, and we'll discuss how these are considered separate roles in a bit. Let's go left down the pipe towards data analytics. The first step down the pipe towards data analytics is to store what we've already cleaned up in the data in something called a data warehouse. A data warehouse, a data warehouse is separate from a data lake. It's a little different. Typically, what a data warehouse will do is take a slice of historical data from the recent past, transform that data to be represented in column format rather than row format, and then maybe apply some additional transformations on the data. So let's take this step by step. The way you would think about a data warehouse is that it is something like a proxy that sits between the data lake and the data analyst. A proxy, almost like a cache layer between the data source. And analyst in the same way that if you've ever worked with web development, there's a thing called mem cache D, that you can stand up in front of your MySQL server and have it proxy requests, which are frequently accessed. It caches SQL queries. So if there's a lot of really common SQL queries that get called frequently and would put a lot of strain on the underlying database, memc D can sit between the database and the server and cache those common queries so as to reduce strain on the database and speed up SQL queries. A data warehouse is similar to that. We take a slice of data over, let's say the last one year or two years. We transform that data to be represented in column structured format rather than row format. Row format is what you're used to when you're looking at data like a spreadsheet or a database. Now imagine taking the rows and columns and just flipping them around. That's column structured data. That structure of data is more desirable for analysts than for other roles in the data science pipeline, like machine learning engineers. Machine learning models do like row structured data, the type we're familiar with, but data analysts prefer column structured data, and that's because running aggregation queries specifically over columns is faster in that way, like count and sum and average running those queries is faster on column structured data. And then we might cache specific queries, which are frequently ran by analysts, and all these things combined turn a data warehouse into a powerhouse proxy between the data lake, so that very fast cached realtime queries can be run against our data by the analysts. So that's our data warehouse. And then finally, our data pipes out of the data warehouse to the data analysts and the business intelligence professionals or business analysts, bi bi for business intelligence. These are two very similar roles, very related roles, but there is a nuanced difference. And these roles are more likely to be differentiated from each other at larger companies where we have lots of roles to fill as opposed to smaller companies where one person wears many hats. So both of these roles deal in analytics, deal in charts and graphs. We take our data outta the data warehouse and we rotate it around in our hands. We look at the top, we look at the bottom in order to make human decisions. That's what makes these roles different than machine learning. Machine learning makes automated decisions. In other words, data analytics or business intelligence is for humans, and machine learning is for robots. But we'll get to that in a bit. A data analyst is more of a technical role. A data analyst is actually likely to be involved in the entire data science pipeline. Up until this point, everything we've already discussed, such as ingestion into a data lake, feature engineering, and working with a data warehouse is likely to also be tasks performed by a data analyst. In addition to the actual analytics task of charting and graphing, looking at the data in order to come up with human decisions. Whereas a business intelligence person, a bi professional, is typically a less technical role. They're less likely to be involved in the entire data science pipeline up until this point, and instead is likely to be working with the data from the data warehouse as a consumer, as a user. And the giveaway difference between these two roles is the tooling. BI people will be using bi specific tools, things like Tableau, T-A-B-L-E-A-U, and Power BI by Microsoft and QuickSight on Amazon AWS. Whereas a data analyst is probably gonna actually be doing their analytics themselves with code by way of Python in Jupyter notebooks with map plot lib and Seaborn. We'll talk about those things in a future episode or during the data engineering phase of the pipeline by way of things like Amazon Data Wrangler or Amazon Glue Data Brew. So an analyst is more of a technical role and a BI person is a less technical role. That's data analytics. Data driven decision making for humans. So a data analyst may take the tweets out of our data warehouse and then chart and graph some things. Look at the optimal time to tweet during the day and what day of the week based on the engagement of those tweets, number of retweets and number of replies, and then go running to the boss and say, boss, boss, we should be tweeting at noon on Tuesday based on my research of the data. That's a data analyst machine learning. Now we go down the right side of the fork in the road in the branch of our pipes. Machine learning is automated. Predictions, it's predictions and decisions made by robots. Now, when I say robots, of course I'm being colloquial by machine learning models. So rather than charting and graphing data as an analyst would do, a machine learning engineer will design a model using machine learning tooling. Things like TensorFlow, KR os, PyTorch, psychic Learn, and XG Boost, maybe all hosted in the cloud using SageMaker so that it will learn the patterns of the tweets so that it can automate predictions or maybe even take actions. So rather than coming up with the conclusion that it is optimal to send tweets at noon on Tuesday, the machine learning engineer might build a model that literally sends tweets at noon on Tuesday. Hey, take that a step further, and maybe they might write an NLP model whereby the marketing team can feed it something they want to say. The machine learning model will construct the optimal way to say that using NLP. Put that into a tweet. Send it out at noon on Tuesday. So that's the data science pipeline. It is the flow from left to right of data at ingestion. Put that into a lake, apply some stuff upon that lake, whether it's permissioning or a unified schema. Pull it outta the lake and feature engineer it. Clean up the data, store it in a feature store. Veer left towards the analysts first. We hit the data warehouse, which does some further transformation of our data so that it can be a real time proxy to the analysts for fast querying. The analysts have more of a technical role in doing analytics, charting and graphing on data for human decisions, and the BI people have a less technical role for the same veer right to the machine learning engineers who build automative predictive models. Now, what is a data scientist then? Well, that's a little bit nebulous. And the reason is, and the reason is it depends on the size of the company you work for. If you work for a mom and pop shop and you are their only data person, then as a data scientist, your role is everything. Everything I just listed, you're gonna handle each phase in the data pipeline. But if you apply to a larger organization like a Fang Corporation, Facebook, Amazon, apple, Netflix, Google, then it is likely the case that you'll be applying as a role for any one phase of this pipeline. So for example, Google might have an entire data science team, and that team is broken down into the data engineering team, the analytics team, and the machine learning team. And then you might apply to the machine learning team or the feature engineering team. In other words, depending on the size of the company. Any one phase in the hierarchical breakdown of the data science stack is a candidate for a role as a data scientist. And so what you might find is that people may market themselves either as specific roles like data analysts, business intelligence, machine learning engineering, or data engineering, or they may market themselves as a data scientist, quote unquote, either implying that they can handle everything in the stack or that they're looking to be more generally applicable. Maybe they're targeting smaller companies and want to wear multiple hats. And then a company might be looking for a data scientist, typically, meaning they want somebody who can handle any and all of the moving parts in the data science pipeline. Or they may specify which specific role they want, a machine learning engineer, a data analyst, or a feature engineer. And that's why it can be confusing sometimes the difference between machine learning and data science. Now, which of these roles should you as a budding engineer in the space, target, machine learning, data analytics, data engineering, or the whole data science stack? Well take an analogy from web development. There's front end server database and DevOps as an example, and somebody who tackles the entire stack calls themselves a full stack engineer. You could either specialize in front end and get really good at it, and there's a lot to know about just front end, for example. Similarly, machine learning is enough to tackle all on its own and keep you busy for a lifetime. There's ML ops or machine learning operations deploying your machine learning models to the cloud. Developing and training your models on various toolkits like Pie, torch, and TensorFlow. Deploying, optimizing and monitoring your models. All these things could keep you busy for a lifetime, and you could dedicate yourself exclusively to machine learning. That would be the equivalent of somebody dedicating themselves exclusively to front end engineering, where other roles in the data science stack are analogous to other roles in a web development stack. So it depends on if you want to tackle the entire stack at large, or if any one portion of the stack sounds more appealing than another. What about machine learning versus ai? Now, I did mention that machine learning is subsuming the various subdomains of ai, but that's not necessarily or completely the case, at least not yet. And so machine learning is more of an industry practice, a professional field. If you want to get into industry, I would choose machine learning over artificial intelligence. You might get a master's in machine learning or data science, whereas targeting artificial intelligence is kind of that dreamy scientific, chasing the rainbow to the pot of gold of the future. That's more of an academic endeavor. So if you wanna stay in academia and push the scientific frontier, then targeting AI conceptually would be more aligned than targeting machine learning. So that's data science, artificial intelligence, and machine learning. Now I'm going to take the history of AI that I had originally recorded in 20 seventeens episode two, and I'm just gonna insert it here 'cause I don't wanna rerecord that stuff. So if you've already listened to episode two, you can stop listening now. If this is your first listen through. Keep on listening. Okay, so we've covered the definition of artificial intelligence, the definition of machine learning, and now finally, let's go into the history of artificial intelligence. I'm gonna recommend you some really good resources that I've picked up from the, from around the web, some common recommendations by people who are looking for good history information on artificial intelligence. So artificial intelligence goes way, way back, way back further than you imagine starting with Greek mythology. And then, you know, coming out in Jewish mythology with Golems, and this is actually a very. Interesting point to me that artificial intelligence has sort of been conceived of in the dawn of humanity. It's almost like, it's part of, it was kind of our, our quest. And we're gonna get back to that actually in the next episode, which is artificial intelligence inspiration. We'll try to inspire you around artificial intelligence and machine learning right now. Just keep that in mind. We, we've been thinking about artificial intelligence since at least Greek mythology. The first attempt at actually implementing, uh, an automaton, I think this guy's name is Raymond Lowell. In the 13th century, Leonardo da Vinci was working on what? He made some walking animals, automata, Decart, and Leni. You know, they had a lot of kind of. Philosophical musings and unconsciousness and liveness on co-created calculus stuff that was used in the development of AI algorithms and mathematics. And both of them were thinking about ai. So this, you know, theory about AI has been around for a very long time. The real nitty gritty stuff kind of started happening around the 17 hundreds and the 18 hundreds, the period of sort of statistics and mathematical decision making. One of the first most important figures is Thomas Bayes. He's worked on reasoning about the probability of events. So Thomas Bayes reasoning about the probability of events. Remember that name? He's gonna come up again multiple times in the future. Bayesian inference. Is very important fundamental component of machine learning. So he, he's a very big figure. George Bull, who was involved in logical reasoning and binary algebra, got lobbed rege with propositional logic. So these were components in the development of computers in general, but also in the development of machine learning. Charles Babbage and Ada Byron slash Love Lovelace in 1832 designed the analytical engine. Which was a programmable calculating machine. And then it was in 1936 that we got the Universal Turing machine. And I think that a lot of people consider that kind of a kickoff point for computers. Alan Turing designed the con, the concept of a programmable computer with his universal Turing machine. It also just so happens that Alan Turing was very interested in ai. He talked about, he has an article called, UH, computing Machinery and Intelligence. So he was thinking about AI pretty early on as well. 1946, John Vaughn Neuman uses Alan Touring's Universal Touring Machine in the development of his universal computing machine. If I'm not mistaken, I believe the Universal Turing machine was the theoretical and the universal computing machine. He actually made a computer. I think John Numan made the architecture for the first kind of universally programmable computer. And now we finally start to get into the actual machine learning stuff, machine learning and artificial intelligence in 1943. Warren McCulloch and Walter Pitts, they kind of built the computer representation of a neuron that was later refined by Frank Rosenblatt to create the perceptron and that word. The perceptron will come back, come up later. That's the first artificial neuron. And of course you stack those in an artificial neuron by way of what's called a multilayer perceptron, and you've got an artificial neural network, and that is deep learning. That's the fun stuff. So McCulloch and pits and rosenblatt. Three very important figures. It wasn't until 1956 that the word artificial intelligence was coined by John McCarthy at the Dartmouth workshop. So Dar, John McCarthy, Marvin Minsky, Arthur Samuel Oliver Selfridge, Ray Solomonoff, Alan Newell, and Herbert Simon. They all came together at Dartmouth into at this workshop in 1956, and their goal was to simulate all aspects of intelligence, end quote, and that is the definition of artificial intelligence at the beginning of this podcast. So these guys defined ai, they created a field of it, and then they set off working. So at the workshop. It was kind of like a hackathon, like a multi-day hackathon. They just cracked at it. Newell and Simon created heuristics, sort of the branch of heuristics in machine learning created this thing called the general problem solver. Okay. Here we see the creation of computer vision, natural language processing, shaky the robot. So this is between the 1950s and the 1970s. So they were working a little bit on, on this at the actual Dartmouth workshop. But then they all took, took their projects home to their own universities and they continued cracking at these things. We call this sort of the golden era of artificial intelligence. It's when it was like an explosion of interest and, and research and joy and light in the field. Fi feigenbaum creates expert systems. This is also sort of the creation of what we call. Good old fashioned AI because later you'll see we kind of switched gears on the way we do AI in today's generation, good old fashioned AI or gofi, G-O-F-A-I uses an approach called symbolism. These are logic-based approaches, knowledge-based expert systems, basically where the knowledge and the logic is kind of hard coded into the system. It's a little bit, it's a little bit like that once and twice removed analogy I used back when playing checkers. A little bit less magical maybe than the connectionism approach of neural networks. But some of the experts, including Marvin Minsky during this time, took the side of gofi of symbolism and they said that the connectionism approach of artificial neural networks wouldn't hold It had some problems. So this was the golden era, the fifties and the, the fifties to the seventies, the golden era of an explosion of research and ideas and discoveries and developments. By some of the, the most important and influential names in the history of artificial intelligence that you'll read in any book. Um, but there was also a little bit of adversity between some of the players, and like I said, so, so Connectionism began to be criticized due to something what was called com, a combinatorial explosion. Basically, neural networks were too hard, computationally too hard on the computer to be effective at generalizing solutions going forward. On the heel of that, in the 1970s, came this report by James Lighthill called the Lighthill Report, and this did vast damage to the field of artificial intelligence. So, like I said, this was between the fifties and seventies when this golden era was ensuing in artificial intelligence when the world was just vibrant and excited about ai. And in that excitement, darpa, you know, the military defense and lots of companies invested into these guys, these major players, and created companies around these technologies. There was a lot of hype, a lot of excitement, and under delivery, it was almost like a Silicon Valley. Bubble bust. It was almost like they said, we can do anything in the whole universe because we're simulating intelligence. Anything you could imagine intellectually, we can simulate that, but in fact, of course it's a lit little bit more difficult than that. We're getting closer every day. But they underdelivered in too long of a time. And so people started cutting funds and cutting grants and cutting contracts with contractors. And a lot of that was due to this Light Hill report, which was basically a report on all these kind of negative aspects saying, yes, indeed these guys are biting off more than they can chew and they're under delivering. So that created what we call the AI winter. AI went underground and very few people were funding it anymore. The AI winter lasted from about the seventies until the nineties, and it started to make the comeback. And the reason it made a comeback was that AI finally started to have some practical applications. I'm sure it was kind of these diehards that were hiding in the cave. They knew that it would, that it had real application in the industry. And so they finally started making some practical use of AI without some lofty promises. So for example, advertising and recommender engines, which are two of the most commonly used applications of artificial intelligence in the modern era. The other thing is that the computers got better. So remember that one of the reasons for the AI winter under delivery was due to this combinatorial explosion that these generalization, generalizing algorithms could not perform on the computers of that modern time, but they could perform on the computers. 20 years later, 30 years later, it was in the nineties and the two thousands that AI really started to pick up steam again. And finally the last piece of the puzzle was data. One thing that you're gonna find that I'm gonna teach you in later episodes is that the accuracy of your machine learning algorithms improve the more data you have. And with the internet becoming so popular and data becoming so prevalent all over the internet that we could just scrape web pages, and there was Excel spreadsheets and databases everywhere of this and that, these machine learning algorithms had basically just a gold mine of information to work with. This is what the era of what's called big data. Now you can work with all the data at your fingertips, as you can imagine, big data. So AI finally made a comeback, finally had this AI spring after the AI winter. Another reason I think that AI made a comeback, which I don't hear touted a lot, was actually optimizations to the algorithms. I know that they say that our computers got really fast, so the co computation got more efficient, but so did the algorithms. So for example, in 2006, Jeffrey Hinton optimized the back propagation algorithm, which made artificial neural networks substantially more tractable and put connectionism back on the map. And in fact, I believe it was that that sort of solidified connectionism as a really powerful, commonly used technique going forward, and that gofi and symbolism started to go out of vogue. Finally we, we have the modern era, 2000 2017. Currently, Bloomberg says that 2015 was sort of a whopper for the AI industry, and I don't really know why. I think maybe what happened is just that, just that this graph that's constantly beyond the rise since the nineties has just finally hit a peak. That it's become really popular in the modern era. Now. Almost any technology company is adopting machine learning and artificial intelligence. They're hiring machine learning engineers and data scientists like never before. So we're in an explosion of ai. In fact, there's a little bit of concern that it's gonna become another bubble bust, another AI winter. But there's a lot of people who say that that probably won't happen. That we're actually, we're really doing well with the algorithms and the data. The algorithms are very precise due to the amount of data we have, the computing machines. We can scale these algorithms horizontally across an AWS cluster and all these things. So we're in a very, very good time for machine learning. Very interesting and good time to be alive and see what's happening. Now, there's one company that I want to draw attention to, and this will be my final point on the history of artificial intelligence. The company's name is DeepMind. And DeepMind was acquired by Google, I believe, in 2014. DeepMind at first was kind of just playing games. It was an artificial intelligence system using deep learning approaches like Deep Q Networks in order to play old classic games like Pong, and then eventually old console games, and more recently modern console games like from PlayStation and such. But Google saw potential in the types of things that they were doing with deep learning in, uh, reinforcement learning specifically, and they acquired them. And DeepMind has been putting out a lot. They have just been demolishing the field of machine learning and artificial intelligence. They're probably the most present figure in the news of artificial intelligence today. They put out papers all the time, making big splashes in research and developments. Something I mentioned previously called a differentiable computer, which brought the subfield of knowledge, representation, and reasoning into the domain of machine learning. For example, DeepMind is something to keep a very close eye on. They're doing some really interesting work these days. Okay, so that's a history of artificial intelligence. So we covered what is ai, a definition, what is machine learning compared to ai? What is machine learning compared to data science compared to statistics? Why is ML the sort of essential subfield of ai? And what is the history of AI in brief? If you're interested in the history of AI and the definition of ai, I'm going to provide some resources. There's one book that's commonly recommended, and I'll point that out in the show notes. So go to the show notes, O-C-D-E-V-E-L, oc deel.com, and this will be episode two and get those links if you're interested in the history of machine learning. So in the next episode, I'm going to try to, I hope, inspire you to be interested in the field of AI and in specifically in machine learning. I think that this episode was probably rather boring. The next episode of all the episodes I have planned so far. The next episode I think will be the most fun episode. It's gonna be inspiration around the field of ai. We're gonna talk about the singularity consciousness, automate automation of the entire industry, some really philosophical and crazy stuff. So I hope to see you then.
Comments temporarily disabled because Disqus started showing ads (and rough ones). I'll have to migrate the commenting system.