episode stringlengths 45 100 | text stringlengths 1 528 | timestamp_link stringlengths 56 56 |
|---|---|---|
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | That's right. | https://karpathy.ai/lexicap/0005-large.html#00:27:26.120 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | But aren't you surprised by the beauty of it? | https://karpathy.ai/lexicap/0005-large.html#00:27:27.120 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So not mathematical beauty, but the fact that it works at all or are you criticizing that | https://karpathy.ai/lexicap/0005-large.html#00:27:29.120 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | very beauty, our human desire to interpret, to find our silly interpretations in these | https://karpathy.ai/lexicap/0005-large.html#00:27:38.200 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | constructs? | https://karpathy.ai/lexicap/0005-large.html#00:27:47.880 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Let me ask you this. | https://karpathy.ai/lexicap/0005-large.html#00:27:49.840 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Are you surprised and does it inspire you? | https://karpathy.ai/lexicap/0005-large.html#00:27:51.320 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | How do you feel about the success of a system like AlphaGo at beating the game of Go? | https://karpathy.ai/lexicap/0005-large.html#00:27:57.100 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Using neural networks to estimate the quality of a board and the quality of the position. | https://karpathy.ai/lexicap/0005-large.html#00:28:03.520 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | That is your interpretation, quality of the board. | https://karpathy.ai/lexicap/0005-large.html#00:28:11.600 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Yeah, yes. | https://karpathy.ai/lexicap/0005-large.html#00:28:14.600 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Yeah. | https://karpathy.ai/lexicap/0005-large.html#00:28:15.600 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So it's not our interpretation. | https://karpathy.ai/lexicap/0005-large.html#00:28:16.600 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | The fact is a neural network system, it doesn't matter, a learning system that we don't I | https://karpathy.ai/lexicap/0005-large.html#00:28:20.320 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | think mathematically understand that well, beats the best human player, does something | https://karpathy.ai/lexicap/0005-large.html#00:28:25.920 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | that was thought impossible. | https://karpathy.ai/lexicap/0005-large.html#00:28:30.160 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | That means that it's not a very difficult problem. | https://karpathy.ai/lexicap/0005-large.html#00:28:31.160 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So you empirically, we've empirically have discovered that this is not a very difficult | https://karpathy.ai/lexicap/0005-large.html#00:28:35.160 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | problem. | https://karpathy.ai/lexicap/0005-large.html#00:28:40.200 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Yeah. | https://karpathy.ai/lexicap/0005-large.html#00:28:41.200 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | It's true. | https://karpathy.ai/lexicap/0005-large.html#00:28:42.200 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So maybe, can't argue. | https://karpathy.ai/lexicap/0005-large.html#00:28:44.080 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So even more I would say that if they use deep learning, it is not the most effective | https://karpathy.ai/lexicap/0005-large.html#00:28:48.720 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | way of learning theory. | https://karpathy.ai/lexicap/0005-large.html#00:28:56.680 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | And usually when people use deep learning, they're using zillions of training data. | https://karpathy.ai/lexicap/0005-large.html#00:29:00.320 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Yeah. | https://karpathy.ai/lexicap/0005-large.html#00:29:08.800 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | But you don't need this. | https://karpathy.ai/lexicap/0005-large.html#00:29:10.480 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So I describe challenge, can we do some problems which do well deep learning method, this deep | https://karpathy.ai/lexicap/0005-large.html#00:29:13.520 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | net, using hundred times less training data. | https://karpathy.ai/lexicap/0005-large.html#00:29:23.240 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Even more, some problems deep learning cannot solve because it's not necessary they create | https://karpathy.ai/lexicap/0005-large.html#00:29:28.400 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | admissible set of function. | https://karpathy.ai/lexicap/0005-large.html#00:29:38.560 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | To create deep architecture means to create admissible set of functions. | https://karpathy.ai/lexicap/0005-large.html#00:29:40.840 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | You cannot say that you're creating good admissible set of functions. | https://karpathy.ai/lexicap/0005-large.html#00:29:45.840 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | You just, it's your fantasy. | https://karpathy.ai/lexicap/0005-large.html#00:29:50.680 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | It does not come from us. | https://karpathy.ai/lexicap/0005-large.html#00:29:52.760 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | But it is possible to create admissible set of functions because you have your training | https://karpathy.ai/lexicap/0005-large.html#00:29:54.960 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | data. | https://karpathy.ai/lexicap/0005-large.html#00:30:00.280 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | That actually for mathematicians, when you consider a variant, you need to use law of | https://karpathy.ai/lexicap/0005-large.html#00:30:01.280 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | large numbers. | https://karpathy.ai/lexicap/0005-large.html#00:30:10.600 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | When you're making training in existing algorithm, you need uniform law of large numbers, which | https://karpathy.ai/lexicap/0005-large.html#00:30:11.600 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | is much more difficult, it requires VC dimension and all this stuff. | https://karpathy.ai/lexicap/0005-large.html#00:30:20.840 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | But nevertheless, if you use both weak and strong way of convergence, you can decrease | https://karpathy.ai/lexicap/0005-large.html#00:30:25.300 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | a lot of training data. | https://karpathy.ai/lexicap/0005-large.html#00:30:33.480 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | You could do the three, the swims like a duck and quacks like a duck. | https://karpathy.ai/lexicap/0005-large.html#00:30:35.240 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So let's step back and think about human intelligence in general. | https://karpathy.ai/lexicap/0005-large.html#00:30:41.360 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Clearly that has evolved in a non mathematical way. | https://karpathy.ai/lexicap/0005-large.html#00:30:48.820 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | It wasn't, as far as we know, God or whoever didn't come up with a model and place in our | https://karpathy.ai/lexicap/0005-large.html#00:30:54.120 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | brain of admissible functions. | https://karpathy.ai/lexicap/0005-large.html#00:31:04.280 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | It kind of evolved. | https://karpathy.ai/lexicap/0005-large.html#00:31:05.880 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | I don't know, maybe you have a view on this. | https://karpathy.ai/lexicap/0005-large.html#00:31:06.880 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So Alan Turing in the 50s, in his paper, asked and rejected the question, can machines think? | https://karpathy.ai/lexicap/0005-large.html#00:31:09.720 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | It's not a very useful question, but can you briefly entertain this useful, useless question? | https://karpathy.ai/lexicap/0005-large.html#00:31:16.920 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Can machines think? | https://karpathy.ai/lexicap/0005-large.html#00:31:23.960 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So talk about intelligence and your view of it. | https://karpathy.ai/lexicap/0005-large.html#00:31:25.720 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | I don't know that. | https://karpathy.ai/lexicap/0005-large.html#00:31:28.560 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | I know that Turing described imitation. | https://karpathy.ai/lexicap/0005-large.html#00:31:29.880 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | If computer can imitate human being, let's call it intelligent. | https://karpathy.ai/lexicap/0005-large.html#00:31:35.560 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | And he understands that it is not thinking computer. | https://karpathy.ai/lexicap/0005-large.html#00:31:43.060 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | He completely understands what he's doing. | https://karpathy.ai/lexicap/0005-large.html#00:31:46.720 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | But he set up problem of imitation. | https://karpathy.ai/lexicap/0005-large.html#00:31:49.480 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So now we understand that the problem is not in imitation. | https://karpathy.ai/lexicap/0005-large.html#00:31:53.840 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | I'm not sure that intelligence is just inside of us. | https://karpathy.ai/lexicap/0005-large.html#00:31:58.000 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | It may be also outside of us. | https://karpathy.ai/lexicap/0005-large.html#00:32:04.360 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | I have several observations. | https://karpathy.ai/lexicap/0005-large.html#00:32:06.680 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So when I prove some theorem, it's very difficult theorem, in couple of years, in several places, | https://karpathy.ai/lexicap/0005-large.html#00:32:09.440 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | people prove the same theorem, say, Sawyer Lemma, after us was done, then another guys | https://karpathy.ai/lexicap/0005-large.html#00:32:20.360 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | proved the same theorem. | https://karpathy.ai/lexicap/0005-large.html#00:32:27.140 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | In the history of science, it's happened all the time. | https://karpathy.ai/lexicap/0005-large.html#00:32:28.960 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | For example, geometry, it's happened simultaneously, first it did Lobachevsky and then Gauss and | https://karpathy.ai/lexicap/0005-large.html#00:32:32.280 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Boyai and another guys, and it's approximately in 10 times period, 10 years period of time. | https://karpathy.ai/lexicap/0005-large.html#00:32:40.600 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | And I saw a lot of examples like that. | https://karpathy.ai/lexicap/0005-large.html#00:32:48.800 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | And many mathematicians think that when they develop something, they develop something | https://karpathy.ai/lexicap/0005-large.html#00:32:51.760 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | in general which affect everybody. | https://karpathy.ai/lexicap/0005-large.html#00:32:57.800 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So maybe our model that intelligence is only inside of us is incorrect. | https://karpathy.ai/lexicap/0005-large.html#00:33:01.600 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | It's our interpretation. | https://karpathy.ai/lexicap/0005-large.html#00:33:07.320 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | It might be there exists some connection with world intelligence. | https://karpathy.ai/lexicap/0005-large.html#00:33:09.320 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | I don't know. | https://karpathy.ai/lexicap/0005-large.html#00:33:15.800 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | You're almost like plugging in into... | https://karpathy.ai/lexicap/0005-large.html#00:33:16.800 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Yeah, exactly. | https://karpathy.ai/lexicap/0005-large.html#00:33:19.040 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | And contributing to this... | https://karpathy.ai/lexicap/0005-large.html#00:33:21.240 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Into a big network. | https://karpathy.ai/lexicap/0005-large.html#00:33:22.640 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Into a big, maybe in your own network. | https://karpathy.ai/lexicap/0005-large.html#00:33:24.360 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | On the flip side of that, maybe you can comment on big O complexity and how you see classifying | https://karpathy.ai/lexicap/0005-large.html#00:33:28.360 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | algorithms by worst case running time in relation to their input. | https://karpathy.ai/lexicap/0005-large.html#00:33:37.400 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So that way of thinking about functions, do you think p equals np, do you think that's | https://karpathy.ai/lexicap/0005-large.html#00:33:42.240 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | an interesting question? | https://karpathy.ai/lexicap/0005-large.html#00:33:47.840 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Yeah, it is an interesting question. | https://karpathy.ai/lexicap/0005-large.html#00:33:49.120 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | But let me talk about complexity in about worst case scenario. | https://karpathy.ai/lexicap/0005-large.html#00:33:52.000 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | There is a mathematical setting. | https://karpathy.ai/lexicap/0005-large.html#00:34:00.000 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | When I came to United States in 1990, people did not know, they did not know statistical | https://karpathy.ai/lexicap/0005-large.html#00:34:04.320 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | learning theory. | https://karpathy.ai/lexicap/0005-large.html#00:34:11.160 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | So in Russia, it was published to monographs, our monographs, but in America they didn't | https://karpathy.ai/lexicap/0005-large.html#00:34:13.040 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | know. | https://karpathy.ai/lexicap/0005-large.html#00:34:19.400 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Then they learned and somebody told me that it is worst case theory and they will create | https://karpathy.ai/lexicap/0005-large.html#00:34:20.400 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | real case theory, but till now it did not. | https://karpathy.ai/lexicap/0005-large.html#00:34:26.640 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | Because it is mathematical too. | https://karpathy.ai/lexicap/0005-large.html#00:34:30.800 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | You can do only what you can do using mathematics. | https://karpathy.ai/lexicap/0005-large.html#00:34:34.100 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | And which has a clear understanding and clear description. | https://karpathy.ai/lexicap/0005-large.html#00:34:38.680 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | And for this reason, we introduce complexity. | https://karpathy.ai/lexicap/0005-large.html#00:34:45.920 |
Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 | And we need this because using, actually it is diversity, I like this one more. | https://karpathy.ai/lexicap/0005-large.html#00:34:52.640 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.