text stringlengths 1 14.3k |
|---|
in meters/second |
I won't lie, my eyes glossed over at that. I just pre-assumed it was at least partial nonsense and moved on. |
you can't divide a weight, by an acceleration, and come up with another weight! |
And 200lbs, at 1G.. is still 200lbs of mass |
Hell, LOOK AT THE EQUASION |
it divided 200lbs by 9.81 and gave the result in kg! |
the thing with chatGPT it that it needs you to put all the data you want it to take in consideration |
something I do is starting by the simplest question and then adding stuff it replies as "would be needed in order to" |
if you want it to take a specific friction and aerodynamics value in consideration, you gotta say it xD |
otherwise it goes himself for some average/estimation which is probably causing that ramp of issues |
my 3rd grade math teacher would give me an F if I did that! |
But it ignored the parameters you *did* give it, to give you a nonsense answer. |
*how to DIY 101* |
Its literally violating basic math unit conversion |
and I don't just mean 'it gave the results in kg' |
I mean, it divided 200lbs, by 9.81 = 20.38 |
and then just said.. Oh, this is a scientific math equasion! that must mean the results are in KG |
and... slapped a KG on there, without converting to kg |
Actually no, I misread the initial question. So ignore what I said about it ignoring input parameters. |
But as Black Moons is saying, nothing it produced is real. It's not "missing parameters", it's just *wrong* from start to finish. It's not doing anything like real math. |
9.81^2 * |
@Grey Earths yea it even ignored the ^2 |
it just completely went Oh, 9.81 is a number that sounds good, I'll put that there! its often used for speed calculations! |
And often these equasions result in KG, So i'll say the result is in KG! |
Yea. ChatGPT is incredible, but it's still just word association. Really advanced word association. |
Literally, its just going "oh, these numbers are often used for this kind of question, So i'll just shit them out in random order" |
Is it an app? |
When it's right, it's amazing. |
But it's *so so wrong* a lot of the time, and you need to be able to spot that, because it won't tell you it might be wrong, ever. |
To trust anything its doing, is like trusting the guy on the street corner who hasn't taken his meds and is raving and ranting about aliens implanting microchips in his brain, to do your homework |
It will literally tell you that your wrong long before it admits its own failure lol |
then people say AIs are gonna take people's jobs |
twitch just fired 400 employees today |
I hope it wasn't to replace them with AI :facepalm: |
It'll admit it's wrong if you straight up tell it that something it said is wrong. |
It will often struggle to incorporate your correction into whatever it's doing for you, though, and that won't stop it from then being wrong about everything else. |
Except try saying so in a less sure tone |
thats the other thing chatgpt does, is ooze confidence |
lol yep |
It assumes anyone who is confident in what they say is correct. |
Since confidence has a high correlation with correctness. |
... Can't wait for it to start prescribing ivermeticin to people because of so many blogs where people are 100% confident it cures all that ails you. |
And then those people will make blog posts about 'how even AI agrees with us', that the AI will then read.. |
for some stuff it's good at making corrections and also adding new ways to output the same info |
big keke in that one |
From what I've read, ChatGPT's "voice" isn't so much the direct result of its training set, but rather the result of weighted user testing prior to wider release. People would use ChatGPT and rate its responses, and that super-confident-about-everything persona is the result of the weights of that feedback. |
Makes sense. |
People who have no clue what they are looking at, are going to rate the 'confident' awnser better then the 'well maybe its this but I dunno, you should prob google that to make sure im not bullshiting you, resulting in you miscalculating this load bearing structure causing thousands of deaths" |
yet I see an audience for that kind of honest answers |
Incidentally, ChatGPT4 is out and is supposedly much less likely to hallucinate in its responses (though by like 60%, not nearly enough to start generally trusting it), but it's only available to paid users for now. |
gonna save me this convo and ask the same again once v4 is available for freebies |
üëçüèΩ |
Nice. I'm curious to hear how that goes. |
'much less likely to hallucinate in its responses' lols. |
You mean when you ask it how much force it needs to make a 200lbs go 25kph, it won't instead tell you how much force is needed to *checks notes* accelerate 20.39kg to 25kph in one second instead? |
'hallucinations' seems to be the settled-upon term for when an AI just makes shit up. lol |
Maybe! |
Or maybe it'll find some new way to misinterpret the prompt, but use something mostly closer to real math to solve it. |
I don't know. I haven't used GPT4. lol |
Sorry. 26.46km/h |
Dunno where the extra 1.46km/h came from. |
lol |
oh, found it |
Honestly though, I find it amazing that it can spew such nonsense for some prompts, but then turn around and write entire functioning scripts when given other prompts. |
Apparently they where trying to calculate... the... acceleration... from... getting the tire up to speed? |
or.. something stupid like that |
but did so by dividing the velocity (and ignoring the ^2 as it did before, because wtf is ^2 mean anyway! best just to ignore it) and then dividing it by 2x2.07m (Somehow, this got transformed into.. 2/2.07, because what is a x symbol anyway!) |
6.94 / (2 / 2.07) = 7.1829... |
... so no |
I can't even follow wtf is doing |
I think the simplest way I can think of to describe GPT's math skills is that it has a pretty decent grasp of what math is supposed to look like, but it has *no idea* how to do math. lol |
It couldn't even get basic multiplying and dividing correct |
It is literally less useful then windows calculator |
https://tenor.com/view/clap-slowclap-notimpressed-gif-8195521 |
xD |
You also can't divide an acceleration by a distance and come up with.. an acceleration. |
THATS NOT HOW UNITS WORK |
@EPHIIOS right? RIGHT? THATS NOT HOW UNITS WORK! |
*In Base 4, I'm fine!* |
Basically, Its failing to even pass grade 3 math from what im seeing here. |
Just randomly multiplying units together with no concept of how that actually works and not even getting the proper result even if you ignore the fact its using improper units all over the place |
Yea. Because, as far as I understand, nobody's ever set out to teach it math. |
It just looked at a lot of numbers on the internet and did its best to draw its own patterns. |
Or rather, if I understand correctly, attempts to draw those patterns on the fly. |
So its someone who failed grade 3 math who thinks they passed university math with flying colors. |
chatgpt is literally modeled after trump. "Nobody knows math better then me, everybody is saying it" |
If you ask it 2+2, it'll probably say 4 every time (unless you tell it not to lol) because that's 2+2 in the vast majority of its training set. But give it complex math that it can't just pull digit-for-digit from its training set, and it quickly reveals that it doesn't actually know math at all. |
Except it'll try to string some nonsense together instead of admitting that it doesn't really know what math is or how it's done. lol |
at least it's honest |
That's actually a really solid answer |
Although I'd love to see the "I can explain how to solve it step-by-step" claim tested with something complicated. lol |
Just based on how well it *does* math, I'd be very surprised if it can consistently *explain* math. |
Considering it didn't explain anything other then 'what the equasions where supposed to look like if I didn't fuck them up...' |
Wait why the hell was Force defined as "Weight of load / Number of wheels" lol |
@NCC1941 And redefined, with the first definition of force never used anywhere |
Sorry, scrolled back up to the original text to see if GPT had tried to *explain* any of what it was doing, and that jumped out at me as a "how the hell didn't I spot that immediately?" thing |
lol yea |
the whole thing is like "yea if you where not paying attention this all looks legit" |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.