episode
stringlengths
45
100
text
stringlengths
1
528
timestamp_link
stringlengths
56
56
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
abstraction, maybe the notion that objects exist. Actually, I think that's pretty important. And a
https://karpathy.ai/lexicap/0015-large.html#00:39:56.880
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
lot of people won't give you that to start with. Right? So almost like a convolution in the, uh,
https://karpathy.ai/lexicap/0015-large.html#00:40:02.080
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
uh, in the object, semantic object space or some kind of, some kind of ideas in there.
https://karpathy.ai/lexicap/0015-large.html#00:40:08.960
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
That's right. And people are starting like the graph, graph convolutions are an idea that are
https://karpathy.ai/lexicap/0015-large.html#00:40:13.840
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
related to relation, relational representations. And so, so I think there are, so you, I've come
https://karpathy.ai/lexicap/0015-large.html#00:40:17.760
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
I've come far field from perception, but I think, um, I think the thing that's going to make
https://karpathy.ai/lexicap/0015-large.html#00:40:25.840
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
perception that kind of the next step is actually understanding better what it should produce.
https://karpathy.ai/lexicap/0015-large.html#00:40:30.720
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
Right? So what are we going to do with the output of it? Right? It's fine when what we're going to
https://karpathy.ai/lexicap/0015-large.html#00:40:36.800
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
do with the output is steer. It's less clear when we're just trying to make a one integrated
https://karpathy.ai/lexicap/0015-large.html#00:40:40.640
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
intelligent agent, what should the output of perception be? We have no idea. And how should
https://karpathy.ai/lexicap/0015-large.html#00:40:46.960
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
that hook up to the other stuff? We don't know. So I think the pressing question is,
https://karpathy.ai/lexicap/0015-large.html#00:40:52.560
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
what kinds of structure can we build in that are like the moral equivalent of convolution
https://karpathy.ai/lexicap/0015-large.html#00:40:59.040
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
that will make a really awesome superstructure that then learning can kind of progress on
https://karpathy.ai/lexicap/0015-large.html#00:41:03.520
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
efficiently. I agree. Very compelling description of actually where we stand with the perception
https://karpathy.ai/lexicap/0015-large.html#00:41:09.440
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
problem. You're teaching a course on embodied intelligence. What do you think it takes to
https://karpathy.ai/lexicap/0015-large.html#00:41:13.840
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
build a robot with human level intelligence? I don't know if we knew we would do it.
https://karpathy.ai/lexicap/0015-large.html#00:41:19.120
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
If you were to, I mean, okay. So do you think a robot needs to have a self awareness,
https://karpathy.ai/lexicap/0015-large.html#00:41:27.680
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
consciousness, fear of mortality, or is it, is it simpler than that? Or is consciousness a simple
https://karpathy.ai/lexicap/0015-large.html#00:41:36.000
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
thing? Like, do you think about these notions? I don't think much about consciousness. Even
https://karpathy.ai/lexicap/0015-large.html#00:41:44.160
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
most philosophers who care about it will give you that you could have robots that are zombies,
https://karpathy.ai/lexicap/0015-large.html#00:41:50.880
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
right? That behave like humans, but are not conscious. And I, at this moment would be happy
https://karpathy.ai/lexicap/0015-large.html#00:41:55.840
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
enough with that. So I'm not really worried one way or the other. So the technical side,
https://karpathy.ai/lexicap/0015-large.html#00:42:00.320
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
you're not thinking of the use of self awareness. Well, but I, okay, but then what does self
https://karpathy.ai/lexicap/0015-large.html#00:42:03.760
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
awareness mean? I mean, that you need to have some part of the system that can observe other
https://karpathy.ai/lexicap/0015-large.html#00:42:09.920
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
parts of the system and tell whether they're working well or not. That seems critical.
https://karpathy.ai/lexicap/0015-large.html#00:42:16.960
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
So does that count as, I mean, does that count as self awareness or not? Well, it depends on whether
https://karpathy.ai/lexicap/0015-large.html#00:42:21.200
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
you think that there's somebody at home who can articulate whether they're self aware. But clearly,
https://karpathy.ai/lexicap/0015-large.html#00:42:27.360
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
if I have like, you know, some piece of code that's counting how many times this procedure gets
https://karpathy.ai/lexicap/0015-large.html#00:42:33.120
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
executed, that's a kind of self awareness, right? So there's a big spectrum. It's clear you have to
https://karpathy.ai/lexicap/0015-large.html#00:42:37.600
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
have some of it. Right. You know, we're quite far away in many dimensions, but is there a direction
https://karpathy.ai/lexicap/0015-large.html#00:42:43.680
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
of research that's most compelling to you for, you know, trying to achieve human level intelligence
https://karpathy.ai/lexicap/0015-large.html#00:42:48.160
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
in our robots? Well, to me, I guess the thing that seems most compelling to me at the moment is this
https://karpathy.ai/lexicap/0015-large.html#00:42:54.720
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
question of what to build in and what to learn. Um, I think we're, we don't, we're missing a bunch
https://karpathy.ai/lexicap/0015-large.html#00:43:00.880
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
of ideas and, and we, you know, people, you know, don't you dare ask me how many years it's going to
https://karpathy.ai/lexicap/0015-large.html#00:43:10.160
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
be until that happens because I won't even participate in the conversation because I think
https://karpathy.ai/lexicap/0015-large.html#00:43:17.200
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
we're missing ideas and I don't know how long it's going to take to find them. So I won't ask you how
https://karpathy.ai/lexicap/0015-large.html#00:43:22.320
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
many years, but, uh, maybe I'll ask you what it, when you'll be sufficiently impressed that we've
https://karpathy.ai/lexicap/0015-large.html#00:43:26.400
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
achieved it. So what's, what's a good test of intelligence? Do you like the Turing test, the
https://karpathy.ai/lexicap/0015-large.html#00:43:34.240
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
natural language in the robotic space? Is there something where you would sit back and think,
https://karpathy.ai/lexicap/0015-large.html#00:43:40.080
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
Oh, that's, that's pretty impressive. Uh, as a test, as a benchmark, do you think about these
https://karpathy.ai/lexicap/0015-large.html#00:43:46.400
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
kinds of problems? No, I resist. I mean, I think all the time that we spend arguing about those
https://karpathy.ai/lexicap/0015-large.html#00:43:52.000
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
kinds of things could be better spent just making the robots work better. Uh, so you don't value
https://karpathy.ai/lexicap/0015-large.html#00:43:57.760
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
competition. So, I mean, there's a nature of benchmark benchmarks and datasets or Turing
https://karpathy.ai/lexicap/0015-large.html#00:44:03.520
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
test challenges where everybody kind of gets together and tries to build a better robot
https://karpathy.ai/lexicap/0015-large.html#00:44:10.000
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
cause they want to out compete each other. Like the DARPA challenge with the autonomous vehicles.
https://karpathy.ai/lexicap/0015-large.html#00:44:14.240
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
Do you see the value of that or it can get in the way? I think it can get in the way. I mean,
https://karpathy.ai/lexicap/0015-large.html#00:44:18.640
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
some people, many people find it motivating. And so that's good. I find it anti motivating
https://karpathy.ai/lexicap/0015-large.html#00:44:25.040
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
personally. Uh, but I think what, I mean, I think you get an interesting cycle where for a contest,
https://karpathy.ai/lexicap/0015-large.html#00:44:29.520
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
a bunch of smart people get super motivated and they hack their brains out and much of what gets
https://karpathy.ai/lexicap/0015-large.html#00:44:37.440
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
done is just hacks, but sometimes really cool ideas emerge. And then that gives us something
https://karpathy.ai/lexicap/0015-large.html#00:44:42.000
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
to chew on after that. So I'm, it's not a thing for me, but I don't, I don't regret that other
https://karpathy.ai/lexicap/0015-large.html#00:44:47.200
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
people do it. Yeah. It's like you said with everything else that it makes us good. So jumping
https://karpathy.ai/lexicap/0015-large.html#00:44:54.400
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
topics a little bit, you started the journal of machine learning research and served as its editor
https://karpathy.ai/lexicap/0015-large.html#00:44:59.120
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
in chief. Uh, how did the publication come about and what do you think about the current publishing
https://karpathy.ai/lexicap/0015-large.html#00:45:05.440
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
model space in machine learning artificial intelligence? Okay, good. So it came about
https://karpathy.ai/lexicap/0015-large.html#00:45:13.760
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
because there was a journal called machine learning, which still exists, which was owned by
https://karpathy.ai/lexicap/0015-large.html#00:45:19.680
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
Cluer and there was, I was on the editorial board and we used to have these meetings annually where
https://karpathy.ai/lexicap/0015-large.html#00:45:24.000
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
we would complain to Cluer that it was too expensive for the libraries and that people
https://karpathy.ai/lexicap/0015-large.html#00:45:30.880
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
couldn't publish. And we would really like to have some kind of relief on those fronts and they would
https://karpathy.ai/lexicap/0015-large.html#00:45:34.640
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
always sympathize, but not do anything. So, uh, we just decided to make a new journal and, uh,
https://karpathy.ai/lexicap/0015-large.html#00:45:39.200
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
there was the journal of AI research, which has, was on the same model, which had been in existence
https://karpathy.ai/lexicap/0015-large.html#00:45:46.960
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
for maybe five years or so, and it was going on pretty well. So, uh, we just made a new journal.
https://karpathy.ai/lexicap/0015-large.html#00:45:52.720
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
It wasn't, I mean, um, I don't know, I guess it was work, but it wasn't that hard. So basically
https://karpathy.ai/lexicap/0015-large.html#00:45:59.920
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
the editorial board, probably 75% of the editorial board of, uh, machine learning resigned and we
https://karpathy.ai/lexicap/0015-large.html#00:46:05.280
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
founded the new journal, but it was sort of, it was more open. Yeah. Right. So it's completely
https://karpathy.ai/lexicap/0015-large.html#00:46:14.560
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
open. It's open access. Actually, uh, uh, I had a postdoc, George Conidaris who wanted to call
https://karpathy.ai/lexicap/0015-large.html#00:46:21.760
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
these journals free for all, uh, because there were, I mean, it both has no page charges and has
https://karpathy.ai/lexicap/0015-large.html#00:46:28.960
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
no, uh, uh, access restrictions. And the reason, and so lots of people, I mean, there were, there
https://karpathy.ai/lexicap/0015-large.html#00:46:36.240
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
were people who were mad about the existence of this journal who thought it was a fraud or
https://karpathy.ai/lexicap/0015-large.html#00:46:44.640
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
something. It would be impossible. They said to run a journal like this with basically, I mean,
https://karpathy.ai/lexicap/0015-large.html#00:46:48.960
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
for a long time, I didn't even have a bank account. Uh, I paid for the lawyer to incorporate and the
https://karpathy.ai/lexicap/0015-large.html#00:46:54.320
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
IP address and it just did cost a couple of hundred dollars a year to run. It's a little bit
https://karpathy.ai/lexicap/0015-large.html#00:47:00.320
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
more now, but not that much more, but that's because I think computer scientists are competent
https://karpathy.ai/lexicap/0015-large.html#00:47:06.960
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
and autonomous in a way that many scientists and other fields aren't. I mean, at doing these kinds
https://karpathy.ai/lexicap/0015-large.html#00:47:13.920
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
of things, we already types out our own papers. We all have students and people who can hack a
https://karpathy.ai/lexicap/0015-large.html#00:47:19.440
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
website together in an afternoon. So the infrastructure for us was like, not a problem,
https://karpathy.ai/lexicap/0015-large.html#00:47:24.480
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
but for other people in other fields, it's a harder thing to do. Yeah. And this kind of
https://karpathy.ai/lexicap/0015-large.html#00:47:29.280
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
open access journal is nevertheless one of the most prestigious journals. So it's not like, uh,
https://karpathy.ai/lexicap/0015-large.html#00:47:34.320
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
prestige and it can be achieved without any of the... Paper is not required for prestige.
https://karpathy.ai/lexicap/0015-large.html#00:47:41.600
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
Yeah. It turns out. Yeah. So on the review process side of actually a long time ago,
https://karpathy.ai/lexicap/0015-large.html#00:47:47.520
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
I don't remember when I reviewed a paper where you were also a reviewer. And I remember reading
https://karpathy.ai/lexicap/0015-large.html#00:47:53.600
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
your review being influenced by it and it was really well written. It influenced how I write
https://karpathy.ai/lexicap/0015-large.html#00:47:59.440
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
feature reviews. Uh, you disagreed with me actually. Uh, and you made it, uh, my review,
https://karpathy.ai/lexicap/0015-large.html#00:48:04.080
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
but much better. So, but nevertheless, the review process, you know, has its, uh, flaws.
https://karpathy.ai/lexicap/0015-large.html#00:48:11.520
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
And how do you think, what do you think works well? How can it be improved?
https://karpathy.ai/lexicap/0015-large.html#00:48:19.280
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
So actually when I started JMLR, I wanted to do something completely different.
https://karpathy.ai/lexicap/0015-large.html#00:48:23.600
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
And I didn't because it felt like we needed a traditional journal of record. And so we just
https://karpathy.ai/lexicap/0015-large.html#00:48:28.720
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
made JMLR be almost like a normal journal, except for the open access parts of it, basically. Um,
https://karpathy.ai/lexicap/0015-large.html#00:48:34.800
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
increasingly of course, publication is not even a sensible word. You can publish something by
https://karpathy.ai/lexicap/0015-large.html#00:48:43.200
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
putting it in an archive so I can publish everything tomorrow. So making stuff public
https://karpathy.ai/lexicap/0015-large.html#00:48:47.600
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
is, there's no barrier. We still need curation and evaluation. I don't have time to read all
https://karpathy.ai/lexicap/0015-large.html#00:48:54.240
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
of archive. And you could argue that kind of social thumbs upping of articles suffices,
https://karpathy.ai/lexicap/0015-large.html#00:49:04.560
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
right? You might say, Oh, heck with this. We don't need journals at all. We'll put everything
https://karpathy.ai/lexicap/0015-large.html#00:49:20.000
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
on archive and people will upvote and downvote the articles. And then your CV will say, Oh man,
https://karpathy.ai/lexicap/0015-large.html#00:49:24.880
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
he got a lot upvotes. So, uh, that's good. Um, but I think there's still
https://karpathy.ai/lexicap/0015-large.html#00:49:29.840
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
value in careful reading and commentary of things. And it's hard to tell when people are
https://karpathy.ai/lexicap/0015-large.html#00:49:39.040
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
upvoting and downvoting or arguing about your paper on Twitter and Reddit, whether they know
https://karpathy.ai/lexicap/0015-large.html#00:49:46.320
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
what they're talking about, right? So then I have the second order problem of trying to decide whose
https://karpathy.ai/lexicap/0015-large.html#00:49:53.280
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
opinions I should value and such. So I don't know what I w if I had infinite time, which I don't,
https://karpathy.ai/lexicap/0015-large.html#00:49:57.760
Leslie Kaelbling: Reinforcement Learning, Planning, and Robotics | Lex Fridman Podcast #15
and I'm not going to do this because I really want to make robots work. But if I felt inclined to do
https://karpathy.ai/lexicap/0015-large.html#00:50:04.480