input
stringlengths 0
513
| output
stringlengths 0
513
|
|---|---|
start
|
I might have to tell all of you to drop all of your Wi-fi signals in here. Just so it doesn't crash this tablet.
|
I might have to tell all of you to drop all of your Wi-fi signals in here. Just so it doesn't crash this tablet.
|
But anyway.
|
But anyway.
|
should be good for now, just for a few more
|
should be good for now, just for a few more
|
gaps
|
gaps
|
here. Okay?
|
here. Okay?
|
And you can use these types of models to generate like synthetic images which we've talked about a bit from before
|
And you can use these types of models to generate like synthetic images which we've talked about a bit from before
|
they can actually generate some pretty nice ones. These days. People are using diffusion models more than Gans, because the diffusion models are just way, less of a headache to be able to train, and they tend to actually map these distributions much nicer.
|
they can actually generate some pretty nice ones. These days. People are using diffusion models more than Gans, because the diffusion models are just way, less of a headache to be able to train, and they tend to actually map these distributions much nicer.
|
So kind of like. How I was saying, if we train a generative model, and I used to pick up
|
So kind of like. How I was saying, if we train a generative model, and I used to pick up
|
on like the average data in the middle, a lot more Gans are pretty guilty of that problem. For sure, the fusion models are a little bit better.
|
on like the average data in the middle, a lot more Gans are pretty guilty of that problem. For sure, the fusion models are a little bit better.
|
and this is where we can at least take
|
and this is where we can at least take
|
now these models that are generating images. Try to plug in a little piece to it that says, Please now generate something in the style of this image.
|
now these models that are generating images. Try to plug in a little piece to it that says, Please now generate something in the style of this image.
|
and that'll work pretty well.
|
and that'll work pretty well.
|
and you can do this for a whole bunch of different styles. So instead of just doing
|
and you can do this for a whole bunch of different styles. So instead of just doing
|
translation, you can try and style it such that it looks like it's in the night. Maybe it's a picture in the winter, or something like that.
|
translation, you can try and style it such that it looks like it's in the night. Maybe it's a picture in the winter, or something like that.
|
So you can generate a lot more data.
|
So you can generate a lot more data.
|
Yeah, like a summer winter switch.
|
Yeah, like a summer winter switch.
|
And then there's a recent piece of work which hopefully, I'll be able to get to.
|
And then there's a recent piece of work which hopefully, I'll be able to get to.
|
Here we go
|
Here we go
|
From a previous student of mine where you can use now like text to image diffusion models.
|
From a previous student of mine where you can use now like text to image diffusion models.
|
So if it's got a diffusion model for generating images.
|
So if it's got a diffusion model for generating images.
|
now, you can feed in some text such that you can identify. Okay, this is of class
|
now, you can feed in some text such that you can identify. Okay, this is of class
|
train. Now please generate me a bunch of images that are essentially of the same class.
|
train. Now please generate me a bunch of images that are essentially of the same class.
|
because maybe you add a bunch of stuff in the data set
|
because maybe you add a bunch of stuff in the data set
|
of a bunch of other metallic things that travel on wheels. In some way, you can figure how to generate some stuff from this distribution.
|
of a bunch of other metallic things that travel on wheels. In some way, you can figure how to generate some stuff from this distribution.
|
So essentially, this is figuring out how to key in. All right. If you've got a particular class, it's trying to share information across classes such that it can generate some more data in that particular class. So interesting type of data augmentation. You can go have a look at this
|
So essentially, this is figuring out how to key in. All right. If you've got a particular class, it's trying to share information across classes such that it can generate some more data in that particular class. So interesting type of data augmentation. You can go have a look at this
|
paper. If you want to know more about this particular model.
|
paper. If you want to know more about this particular model.
|
it is a bit tricky to actually train.
|
it is a bit tricky to actually train.
|
He shares his final chicken version
|
He shares his final chicken version
|
like.
|
like.
|
no, okay.
|
no, okay.
|
I'll send a message.
|
I'll send a message.
|
and I'll update the version of this right after. Yeah, I might have added this on the weekend
|
and I'll update the version of this right after. Yeah, I might have added this on the weekend
|
and forgot to upload this one.
|
and forgot to upload this one.
|
Otherwise you can just have a look for effective data augmentation with the fusion models
|
Otherwise you can just have a look for effective data augmentation with the fusion models
|
from Brandon Chabuco.
|
from Brandon Chabuco.
|
Otherwise, if I don't post the updated slides by tonight. Someone message me on fiance
|
Otherwise, if I don't post the updated slides by tonight. Someone message me on fiance
|
and I'll make sure to put them up.
|
and I'll make sure to put them up.
|
Okay?
|
Okay?
|
So those were a lot of methods for being able to do image data augmentation.
|
So those were a lot of methods for being able to do image data augmentation.
|
Like, I said, there's always a few challenges for being able to do this for tabular data. It's just not as continuous. So our data distributions are a bit more tricky to work with.
|
Like, I said, there's always a few challenges for being able to do this for tabular data. It's just not as continuous. So our data distributions are a bit more tricky to work with.
|
So if each point is basically, if each entry in a table.
|
So if each point is basically, if each entry in a table.
|
it's just in our like d dimensional space. Now.
|
it's just in our like d dimensional space. Now.
|
and we want to generate some new points while respecting some boundaries. Indeed!
|
and we want to generate some new points while respecting some boundaries. Indeed!
|
Then we can figure out how to generate some more data.
|
Then we can figure out how to generate some more data.
|
Often this ends up coming out because we know something about the data set. So you can maybe apply Gaussian noise to
|
Often this ends up coming out because we know something about the data set. So you can maybe apply Gaussian noise to
|
generate some more data, but also add some Gaussy noise to the age. But we should know that no one should have the age of Niagara , and no one should be born in.
|
generate some more data, but also add some Gaussy noise to the age. But we should know that no one should have the age of Niagara , and no one should be born in.
|
or something like that. So usually we can apply the filters
|
or something like that. So usually we can apply the filters
|
on top of the very generic data, and then we can use that rejection sampling method again. So if you accidentally produce one like this, okay, garbage. Generate another one and then keep that part in the data. Set
|
on top of the very generic data, and then we can use that rejection sampling method again. So if you accidentally produce one like this, okay, garbage. Generate another one and then keep that part in the data. Set
|
one more time.
|
one more time.
|
Okay?
|
Okay?
|
And this is where you can use kind of like the classic smoke method does a bit of a version of this sampling in order to get some points there.
|
And this is where you can use kind of like the classic smoke method does a bit of a version of this sampling in order to get some points there.
|
And there's a few more
|
And there's a few more
|
advanced method. If you're getting into this kind of uglier area of having to deal with data sets and lack of data
|
advanced method. If you're getting into this kind of uglier area of having to deal with data sets and lack of data
|
you can use.
|
you can use.
|
I think this is pronounced at a sign. But I'm not really sure, and this will even randomly modify some of the positions of even these generated points.
|
I think this is pronounced at a sign. But I'm not really sure, and this will even randomly modify some of the positions of even these generated points.
|
And there's a few examples of using tabular gans. And actually, there's a couple of good new, like really recent pieces of the research week. Because these things are often text actually using Llms is a good application for figuring how to generate some more tabular data.
|
And there's a few examples of using tabular gans. And actually, there's a couple of good new, like really recent pieces of the research week. Because these things are often text actually using Llms is a good application for figuring how to generate some more tabular data.
|
It's of the few things that works well for text and tabular data.
|
It's of the few things that works well for text and tabular data.
|
Those Lms like typically deep learning, has not been a good thing to apply on tabular data.
|
Those Lms like typically deep learning, has not been a good thing to apply on tabular data.
|
because tabulate data is either too discrete or there's not enough of it.
|
because tabulate data is either too discrete or there's not enough of it.
|
And that's where still often, like somewhat more traditional methods, work the best there.
|
And that's where still often, like somewhat more traditional methods, work the best there.
|
Okay.
|
Okay.
|
Golden Rule.
|
Golden Rule.
|
If you can convince your boss or yourself or your funders to give you some money for more data, that's usually the best way to do things. So go with that if you can.
|
If you can convince your boss or yourself or your funders to give you some money for more data, that's usually the best way to do things. So go with that if you can.
|
And then the last thing yeah, which I'll cover in just a minute
|
And then the last thing yeah, which I'll cover in just a minute
|
is data drift. And this is where you kind of like, it seems like you have multiple classes of data.
|
is data drift. And this is where you kind of like, it seems like you have multiple classes of data.
|
But you actually just are doing something over some sort of data drives.
|
But you actually just are doing something over some sort of data drives.
|
So let's see if you've got this kind of compensation common situation.
|
So let's see if you've got this kind of compensation common situation.
|
your company, you're gonna collect data over the last months.
|
your company, you're gonna collect data over the last months.
|
And then what you're gonna end up deciding to do. You're gonna take the st months of that data as training and validation in the last month as test for some reason. Remember, don't do that. We wanna kind of split things equally so. Everything looks like the same distribution.
|
And then what you're gonna end up deciding to do. You're gonna take the st months of that data as training and validation in the last month as test for some reason. Remember, don't do that. We wanna kind of split things equally so. Everything looks like the same distribution.
|
But it's gonna perform really well on these sets, but very poorly on the test set.
|
But it's gonna perform really well on these sets, but very poorly on the test set.
|
So we want to figure out what's actually going on here.
|
So we want to figure out what's actually going on here.
|
The problem is okay. These weren't part of it didn't end up being part of the same distribution, not necessarily because of anything you did, but
|
The problem is okay. These weren't part of it didn't end up being part of the same distribution, not necessarily because of anything you did, but
|
because we all currently operate on the well, the idea that we exist in linear time.
|
because we all currently operate on the well, the idea that we exist in linear time.
|
and that there are other things that will happen because of that.
|
and that there are other things that will happen because of that.
|
So we can figure out, how can we detect drift and what to do about it. There's a lot of examples for this, you know.
|
So we can figure out, how can we detect drift and what to do about it. There's a lot of examples for this, you know.
|
or whatever. Years ago we'd all be teaching remotely right now, and there's a whole bunch of weird this print distribution during Covid. So if you collected your data during that time, I don't know. Maybe you should throw it out now, and we're all back to normal. Supposedly.
|
or whatever. Years ago we'd all be teaching remotely right now, and there's a whole bunch of weird this print distribution during Covid. So if you collected your data during that time, I don't know. Maybe you should throw it out now, and we're all back to normal. Supposedly.
|
And this will be some indication that our models underperforming.
|
And this will be some indication that our models underperforming.
|
So there is one nice way to be able to do this.
|
So there is one nice way to be able to do this.
|
To do some of these tests is you can do some antagonist validation.
|
To do some of these tests is you can do some antagonist validation.
|
So if you've got your test sets and your training set here, which are colored.
|
So if you've got your test sets and your training set here, which are colored.
|
You can actually like train
|
You can actually like train
|
if a classifier is able to differentiate between the sets. So before we talked about like you compute the mean over something. But now, if I train just a classifier label
|
if a classifier is able to differentiate between the sets. So before we talked about like you compute the mean over something. But now, if I train just a classifier label
|
the stuff from the things in the test set as zeros and the stuff that's in the train set as ones.
|
the stuff from the things in the test set as zeros and the stuff that's in the train set as ones.
|
It should not be able to get this right like it should end up being like wrong. or right, % of the time, because it just can't tell the difference.
|
It should not be able to get this right like it should end up being like wrong. or right, % of the time, because it just can't tell the difference.
|
That's way to do that.
|
That's way to do that.
|
This is more helpful, too. If your data is like not normally distributed. And you still want to figure out if it's in the same distribution.
|
This is more helpful, too. If your data is like not normally distributed. And you still want to figure out if it's in the same distribution.
|
And this is where, yeah, we'll have no drift if we basically can't figure out how to separate this data or some drift if it does exist.
|
And this is where, yeah, we'll have no drift if we basically can't figure out how to separate this data or some drift if it does exist.
|
And this happens a lot like according to weather or whatever. Or there's a lot of periodic functions that happen according to the week.
|
And this happens a lot like according to weather or whatever. Or there's a lot of periodic functions that happen according to the week.
|
Everyone does a lot of stuff or a lot different stuff on the weekend. It's also just a part of the way our society is structured.
|
Everyone does a lot of stuff or a lot different stuff on the weekend. It's also just a part of the way our society is structured.
|
So a lot of kind of odd things happen.
|
So a lot of kind of odd things happen.
|
yeah. And then
|
yeah. And then
|
you end up constructing this data set in some way. Remember, you're not going to look quite at the target. You're gonna look at this test set versus training. Set this distribution.
|
you end up constructing this data set in some way. Remember, you're not going to look quite at the target. You're gonna look at this test set versus training. Set this distribution.
|
And then the results must be compared to some sort of random mixture with some null hypothesis. A classifier will be, you know, worse at differentiating a random mixture that might also be a good check, too.
|
And then the results must be compared to some sort of random mixture with some null hypothesis. A classifier will be, you know, worse at differentiating a random mixture that might also be a good check, too.
|
So like
|
So like
|
this is the nice case we've got like zeros and ones, and different deficits.
|
this is the nice case we've got like zeros and ones, and different deficits.
|
You might also train basically some model that's trying to predict. Just like random stuff. You take all of your labels and just lay random, label them all randomly.
|
You might also train basically some model that's trying to predict. Just like random stuff. You take all of your labels and just lay random, label them all randomly.
|
just to check, as it's kind of a slightly better baseline than just taking %. Maybe your model can somehow guess on a few things for some class, or whatever
|
just to check, as it's kind of a slightly better baseline than just taking %. Maybe your model can somehow guess on a few things for some class, or whatever
|
it's a better. It's an additional way to be able to check this version of looking at different distributions.
|
it's a better. It's an additional way to be able to check this version of looking at different distributions.
|
And hopefully, if you're lucky you'll be in some space that's like a constant drift. No sorry.
|
And hopefully, if you're lucky you'll be in some space that's like a constant drift. No sorry.
|
a constant drift problem, which is also kind of hard. This is sort of
|
End of preview. Expand
in Data Studio
README.md exists but content is empty.
- Downloads last month
- 57