Transcript Intro ML Course of Gatech
Transcript Intro ML Course of Gatech
1 - Definition of ML
Interview Transcript: "Defining Machine Learning"
Michael: Hi Michael.
Charles: Hey Charles, how's it going?
Michael: It's going quite well, how's it going with you?
Charles: Good. Good.
Michael: Good Good. So, today I thought we would talk a little bit about the philosophy of
Machine Learning.
Charles: Oh, I hate philosophy.
Michael: I don't like it much either, although I am a doctor of philosophy.
Charles: Oh, that's very impressive.
Michael: Aren't you a doctor of philosophy too?
Charles: I am, it's kind of impressive.
Michael: It is kind of impressive. So what we wanted to kind of get across today was a little bit
about why we, the class is structured the way it is. What the different parts are. And maybe go a
little bit of back and forth about what we think you should be getting out of the course. That
seem reasonable?
Charles: Sure.
Michael: Okay. Well, so, first off, by the way, before we get started, I wanted to thank you for
coming down to Atlanta, and joining me in these beautiful, studios.
Charles: Well, it's, it's, it's very nice to be here. Thank you for inviting me.
Michael: Oh, no, no, thank you for coming, Michael.
Charles: Thank you for asking me to do the course. This has been a lot of fun.
Michael: Oh, the whole point was to be able to do the course with you, Michael. We like each
other, and that's one of the things that we want you to get, want to get across in this class,
because we like machine learning. We've a lot of stuff in common, but I'm not sure we
completely agree on the most important parts of machine learning and why we do the things
that we do.
Charles: Hm, all right.
Michael: So I think people in the outside world, Michael, would claim that you're more
theoretical than I am.
Charles: In theory.
Michael: In theory, and I'm more practical than you are.
Charles: Practically.
Michael: At least in practice. And hopefully some of that tension will come out in the class. But I
think in order to see why that tension works that way, you have to understand what machine
learning is. So, Michael.
Charles: Right.
Michael: What's machine learning?
Charles: It's about proving theorems.
Michael: [LAUGH] No.
Charles: No?
Michael: I would not say it's about proving theorems, although proving theorems is often
important in machine learning.
Charles: I agree with that.
Michael: Okay. So we're on the same page.
Charles: We're partially on the same page. What is machine learning?
Michael: What is machine learning?
Charles: Give me a definition.
Michael: So it is computational statistics. How's that for a definition?
Charles: That is a definition. It is wrong on so many levels. However, a lot of people would agree
with that statement. They would say that machine learning is really just applied statistics.
Michael: Not applied statistics. Computational statistics.
Charles: Computationally applied statistics. I don't like that definition. I think that it's a bit too
narrow. I think that machine learning is about this broader notion of building artifacts,
computational artifacts, typically. That learn over time based on experience. And then in
particular, it's not just the act of building these artifacts that matter, it's the math behind it. It's
the science behind it. It's the engineering behind it, and it's the computing behind it. It's
everything that goes into building intelligent artifacts that almost by necessity have to learn over
time. You buy that?
Michael: Yeah, so you, you have data, and you do analysis of the data and try to glean things
from the data. And you used, various kinds of computational structure to do that, so,
computational statistics.
Charles: I don't think that's computational statistics.
2 - Supervised Learning
Charles: This class is divided into three subclasses, three parts. They are supervised learning.
Michael: Yeah.
Charles: Unsupervised learning, and reinforcement. So, what do you think supervised learning
is?
Michael: So, I think of supervised learning as being the problem of taking labelled data sets,
gleaning information from it so that you can label new data sets.
Charles: That's fair. I call that function approximation. So, here's an example of supervised
learning. I'm going to give you an input and an output. And I'm going to give them to you as
pairs, and I want you to guess what the function is.
Michael: Sure. Okay?
Charles: Okay. 1, 1.
Michael: Uh-huh.
Charles: 2, 4.
Michael: Wait, hang on, is 1 the input and 1 the output, and 2 the input, and 4 the output?
Charles: Correct.
Michael: Alright. I think I am on to you.
Charles: 3, 9.
Michael: Okay.
Charles: 4, 16.
Michael: Nice.
Charles: 5, 25. 6, 36. 7, 49.
Michael: Nice. This is a very hip data set.
Charles: It is. What's the function?
Michael: It's hip to be squared.
Charles: Exactly. Maybe. So if you believe that's true, then tell me if the input is 10, what's the
output?
Michael: 100.
Charles: And that's right, if it turns out, in fact, that the function is x squared. But the truth is, we
have no idea whether this function is x squared. Not really.
Michael: I have a pretty good idea.
Charles: You do?
Michael: Well-
Charles: Where's that idea come from?
Michael: It comes from having spoken with you over a long period of time. And plus, you know,
math.
Charles: And plus math. Well, I'm going to-
Michael: You can't say I'm wrong.
Charles: You're wrong.
Michael: Oh.
Charles: Yeah, I did.
Michael: You just said I was wrong.
Charles: No, you've talked to me for a long time, and plus math. I agree with that.
Michael: Okay.
Charles: But I'm going to claim that you're making a leap of faith.
Michael: Hm.
Charles: Despite being a scientist, by deciding that the input is 10 and the output is 100.
Michael: Sure. I would agree with that.
Charles: What's that leap of faith?
Michael: Well, I mean, from what you told me, it's still consistent with lots of other mappings
from input to output like 10 gets mapped to 11.
Charles: Right or everything is x squared except 10.
Michael: Sure.
Charles: Or everything is x, x squared up to 10.
Michael: Right, that would be mean-
Charles: That would be mean-
Michael: But it's not logically impossible.
Charles: What would be the median?
Michael: A-ha.
Charles: Thank you very much. I, I was saving that one up.
Charles: Okay, so we've got these three little bits of machine learning here. And there are a lot
of tools and techniques that are inside that.
Michael: Mm-hm.
Charles: And I think that's great. And we're going to be trying to teach you a lot of those tools
and techniques and sort of ways to connect them together. So by the way, as Michael was
pointing out, there are kind of ways that these things might help each other, unsupervised
learning might help supervised learning. It's actually much deeper than that. It turns out you,
even though unsupervised learning is clearly not the same as supervised learning at the level
that we described it, in some ways they're exactly the same thing. Supervised learning you have
some bias. Oh, it's a quadratic function, induction make sense. All these kind of assumptions you
make.
Michael: Mm.
Charles: And in unsupervised learning, I told you that we don't know whether this clusters is
better than this cluster, dividing by sex is better than dividing by height, or, or hair color or
whatever.
Michael: Mm.
Charles: But ultimately you make some decision about how to cluster, and that means implicitly
there's some assume signal. There's some assume set of labels that you think make sense. Oh, I
think things that look alike should somehow be clustered together.
Michael: Mm.
Charles: Or things that are near one another should cluster together. So in some ways it's still
kind of like supervised learning. You can certainly turn any supervised learning problem into an
unsupervised learning problem.
Michael: Mm, mm.
Charles: Right? So in fact, all of these problems are really the same kind of problem.
Michael: Yeah, well there's two things that I'd want to add to that. One is that in some sense, in
many cases, you can formulate all these different problems as some form of optimization. In
supervised learning you want something that, that labels data well and so you're, the thing
you're trying to optimize is find me a function that, that does that scores it. In reinforcement
learning we're trying to find a behavior that scores well. And unsupervised learning we usually
have to make up some kind of a criterion, and then we find a way of clustering the data,
organizing the data so that it scores well. So that was the first point I wanted to make.
Charles: Do you learn about that on the street?
Michael: I learned it in a Math book.
Charles: Yes you [NOISE] I'm, I'm going to move on And so here's the thing.
Michael: Alright.
Charles: Everything just Michael just said except the last part is true. But there's actually a sort
of deeper thing going on here. To me, if you think about the commonalities of everything we've
just said, it boils down to one thing data, data, data, data, data, data. Data is king in machine
learning. Now Michael would call himself a computer scientist.
Michael: Oh, yeah.
Charles: And I would call myself a computationalist.
Michael: What?
Charles: What if I'm in a college of computing at a department of computer science? I believe in
computing and computation as being the ultimate thing. So I would call myself a
computationalist, and Michael would probably agree with that just to keep this discussion
moving.
Michael: Let's say.
Charles: Right. So we're computationalists. We believe in computing. That's a good thing.
Michael: Sure.
Charles: Many of our colleagues who do computations tend to think in terms of algorithms. They
think in terms of what are the series of steps I need to do in order to solve some problem? Or
they.
Michael: [CROSSTALK].
Charles: Might think in terms of theorems. If I try to describe this problem in a particular way, is
it solvable quizzically by some algorithm?
Michael: Yeah.
Charles: And, truthfully, machine learning is a lot of that. But the difference between the person
who's trying to solve our problem as an AI person or as a computing person and somebody
who's trying to solve our problem as a machine learning person is that the algorithm stops being
central, the data starts being central. And so what I hope you get out of this class, or at least
part of the stuff that you do, is understanding that you have to believe the data, you have to do
something with the data, you have to be consistent with the data. The algorithms that fall out of
all that are algorithms, but they're algorithms that take the data as primary or at least
important.
Michael: I'm going to go with co-equal.
Charles: So the algorithms and data are co-equal.
Michael: Well if you believe in Lisp, they're the same thing.
Charles: Exactly!
Michael: Alright. So there you go.
Charles: They knew back in the 70s.
Michael: So it turns out we do agree on most things.
Charles: [NOISE] That was close.
Michael: Excellent! So, the rest of the semester will go exactly like this.
Charles: [LAUGH].
Michael: [LAUGH] Except you won't see us. You'll see our hands though.
Charles: This side. This side.
Michael: You'll see our hands, though. Thank you, Michael.
Charles: It's all right.
Michael: [LAUGH] What?
Charles: [LAUGH] What? [LAUGH].
Michael: That was good, that took me back to when I was four. Okay, so.
Charles: Senor Wences.
Michael: Hm?
Charles: It's called Señor Wences.
Michael: Yes, I know.
Charles: Yeah, okay.
Michael: I'm not that much younger than you are.
Charles: Little bit.
Michael: Ten, 12 years only.
Charles: No come on.
Michael: You can count gray hairs. Anyway the point is the rest of the semester will go like this.
We will talk about supervised learning and a whole series of algorithms. Step back a little bit and
talk about the theory behind them, and try to connect theory of machine learning with theory of
computing notions, or at least that kind of basic idea. What does it mean to be a hard problem
versus an easier problem? Will move into randomized optimization and unsupervised learning
where we will talk about all the issues that we brought up here and try to connect them back to
some of the things that we did in the section on supervised learning. And then finally, we will
spend our time on reinforcement learning. And a generalization of these traditional
reinforcement learning, which involves multiple agents. So we'll talk about a little bit of.
Michael: Mm-hm.
Charles: Game theory, which Michael loves to talk about. I love to talk about. And the
applications of all the stuff that we've been learning to solving problems of how to actually act in
the world? How to build that world out to do something? Or build that agent to play a game or
to teach you how to do whatever you, you need to be taught how to do? But at the end of the
day, we're going to teach you how to think about data, how to think about algorithms, and how
to build artifacts that you know, will learn?
Michael: Let's do this thing.
Charles: Excellent. All right. Well thank you Michael.
Michael: Sure.
Charles: I will see you next time we're in the same place at the same time.