ChatGPT is hilariously incompetent… but on a serious note, I still firmly reject tools like copilot outside demos and the like because they drastically reduce code quality for short term acceleration. That’s a terrible trade-off in terms of cost.
As a software engineer, the number of people I encounter in a given week who either refuse to or are incapable of understanding that distinction baffles and concerns me.
That’s because it’s being advertised as a solution. That’s why you have people worried it’ll take their jobs when in reality it’ll let them do the job better.
The problem I have with it is that all the time it saves me I have to use on reading the code. I probably spend more time on that as once in a while the code it produces is broken in a subtle way.
I see some people swearing by it, which is the opposite of my experience. I suspect that if your coding was copying code from stack overflow then it indeed improved your experience as now this process is streamlined.
I don’t know if it does yet, but if ChatGPT starts providing source for every information, then it would make it much faster to find the relevant information and check their sources, rather than clicking websites one by one.
It used to have a problem with making a claims that were not relevant to or contradicted its own sources, but I don’t recall encountering that problem recently.
Biggest problem with it is that it lies with the exact same confidence it tells the truth.
Or, put another way, it’s confidently incorrect as often as it is confidently correct - and there’s no way to tell the difference unless you already know the answer.
it’s kinda hilarious to me because one of the FIRST things ai researchers did was get models to identify things and output answers together with the confidence of each potential ID, and now we’ve somehow regressed back from that point
i mean giving a confidence for recognizing a certain object in a picture is relatively straightforward.
But LLMs put together words by their likeliness of belonging together under your input (terribly oversimplified).the confidence behind that has no direct relation to how likely the statements made are true. I remember an example where someone made chatgpt say that 2+2 equals 5 because his wife said so. So chatgpt was confident that something is right when the wife says it, simply because it thinks these words to belong together.
Gödel numbers are typically associated with formal mathematical statements, and there isn’t a formal proof for 2+2=5 in standard arithmetic. However, if you’re referring to a non-standard or humorous context, please provide more details.
Of course I don’t know enough about the actual proof for it to be anything but a joke but there are infinite numbers so there should be infinite proofs.
there are also meme proofs out there I assume could be given a Gödel number easily enough.
But LinkedIn bros and corporate people are gonna gobble it up anyways because it has the right buzzwords (including “AI”) and they can squeeze more (low quality) work from devs to brag about how many things they (the corporate owners) are doing.
It’s just a fad. There’s just a small bit that will stay after the hype is gone. You know, like blockchain, AR, metaverse, NFT and whatever it was before that. In a few years there will be another breakthrough with it and we’ll hear from it again for a short while, but for now it’s just a one trick pony.
I should’ve elaborated… It’s supposed to be a vast seamless network of interconnected spaces isn’t it, but that doesn’t exist yet. I think they jumped the gun in that everyone thought Horizons was it. It’s only just starting to be built.
The hilarious thing about blockchain is that the core concept is actively making the whole thing worse. The matrix protocol is sort of essentially blockchain without the decentralized ledger part, and it’s so vastly superior in every single way.
NFTs just show how fundamentally dumb blockchains are, if you skip the decentralized ledger bit then you never need to invent NFT functionality in the first place…
i’m not sure whether matrix uses that, what i’m talking about is how it does the things everyone finds neat about blockchains without the inherent downsides like massive power usage and EVERYONE having to replicate the ENTIRE ledger.
i know IPFS uses merkle trees though, and hilariously blockchains largely rely on that to actually store any significantly sized data.
Well, if you stop listening to people who think it’s a way to get really rich really fast (which it obviously isn’t), cryptocurrencies are quite useful. International transfers are so much cheaper and easier with them.
Yes, you could, for example, use it to manage who is allowed to park in a garage, anonymously. The owner of a parking spot NFT can unlock the door from the outside. Stuff like that.
However, it’s also possible to do that with a small web application. Just payments and transfer of the parking spots are less free and it’s not decentralized.
The proton vpn people are either using or working on using block chain as a sort of email verification. Iirc it won’t have any cost or change in usage to the consumer, just an added layer of security. I’m not smart enough to understand how but it sounds neat.
I may have misread, Its not clear to me if it’ll be a staple of their free email service or will be added as part of their paid subscription to all their services but it still sounds neat. It very well could be bullshit but I’m interested in learning more.
I disagree, because unlike those things, ai actually has a use case. It needs a human supervisor and it isn’t always faster, but chat gpt has been the best educational resource I’ve ever had next to YouTube. It’s also decent at pumping out a lot of lazy work and writing so i don’t have to, or helping me break down a task into smaller parts. As long as you’re not expecting it to solve all your problems for you, it’s an amazing tool.
People said the same things about 3d printing and yeah, while it can’t create literally everything at industrial scale, and it’s not going to see much consumer use, it has found a place in certain applications like prototyping and small scale production.
ChatGPT is OK at summarizing popular, low specificity topics that tons of people have already written a ton about, but it’s terrible at anything else. When I tested its knowledge about the process of a niche interest of mine (fabric dyeing) it skipped completely over certain important pieces of information, and when I prompted it to include them it basically just mirrored my prompt back at me.
Which has pretty much summed up my ChatGPT experience: it just regurgitates stuff I can find myself, but removes the ability to determine if the source is reliable. And if it’s something I’m already having trouble finding detailed information about it usually doesn’t help.
i’d argue that chat gpt is mostly great at taking human bullshit tasks from humans, who dontwant to dothe bullshit, like regurgitating the text from a textbook in different words, writing cover letters for job applications, that are often machine analyzed for buzzwords anyways.
So its use case only exist in the domain of bullshit tasks that only exist to occupy two people without any added value.
Yeah, that is one thing it can do, but it’s not the only thing it does. I’m not sure how to get my point across well but, just because it gives you the wrong answer 25% of the time doesn’t mean it’s useless. In whatever you ask it to do, it often gets you most of the way there as long as you know how to correct it or check its work. The ability to ask specific or follow-up questions when learning something makes it invaluable as a learning tool (if you’re teaching yourself that is (ie. If you’re a university student)). It’s also very useful when brainstorming ideas or helping you approach a problem from a different angle. I can also ask it questions that are far more specific than what a search engine would get me.
It really comes down to if the human operator knows how (and when) to use it properly.
i never trust that information id rather learn with a book videos or just from websites if i ever use it i always fact check it. never blindly trust a computer
That was an integral part to this whole thing, you always fact check the ai. I said this in both of my comments.
I totally get preferring textbooks or videos though. I just find that the ai saves me time since i can ask specific questions about things, and it often gives me concise information that i understand more quickly.
My hairline has started receding very rapidly. There’s there’s these fine hairs all over my desk, and I see the photo I took when joining directly before turning on my camera every meeting.
Doesn’t sood good at all. I’m sorry to hear that, friend. I really hope there’s enough upsides there compared to working at a more mature company for you.
Sort of. Nobody’s cutting corners on aviation structural components, for example. We’ve been pretty good at maximizing general value output, and usually that means lower quality, but not always.
I’m going to say that’s the exception that proves the rule, assuming they were structural parts and not a minor controller chip for de-icing or something.
The company themself announced it without being prompted, and if whoever introduce these unapproved parts into a small number of engines is caught there’s going to be real hell to pay. The stuff that stops you from falling out of the sky is serious business, and is largely treated as such.
On the other hand, a software function that’s hacked together and inefficient will just fly below the radar, and most people will prefer two cheap outfits to one that’s actually well made for the same price, so quality goes right out the window.
I’ve asked ChatGPT to fix simple bugs that were in the code it’s given me. It usually doesn’t understand that there’s a problem and just rewrites it with different variables names.
It’s helped me a bit with resolving weird tomcat/Java issues when upgrading to RHEL8, though. It didn’t give me an answer, but it gave me ideas on where to look (in my case I didn’t realize fapolicyd replaced selinux)
That’s the point - you have the expertise to make proper sense of whatever it outputs. The people pushing for “AI” the most want to rely on it without any necessary expertise or just minimal efforts, like feeding it some of your financial reports and have generate a 5-year strategy only to fail miserably and have no one to blame this time (will still blame anyone else but themselves btw).
It’s not the most useless tool in the world by any means, but the mainstream talk is completely out of touch with reality on the matter, and so are mainstream actions (i.e. overrelying on it and putting way too much faith into it).
An unpopular opinion, I am sure, but if you’re a beginner with something - a new language, a new framework - and hate reading the docs, it’s a great way of just jumping into a new project. Like, I’ve been hacking away on a django web server for a personal project and it saved me a huge amount of time with understanding how apps are structured, how to interact with its settings, registering urls, creating views, the general development lifecycle of the project and the basic commands I need to do what I’m trying to do. God knows Google is a shitshow now and while Stackoverflow is fine and dandy (when it isn’t remarkably toxic and judgmental), the fact is that it cuts down on hours of fruitless research, assuming you’re not asking it to do anything genuinely novel or hyper-specific.
It helps a complete newbie like me get started and even learn while I do. Due to its restrictions and shortcoming, I’ve been having to learn how to structure and plan a project more carefully and thoughtfully, even creating design specs for programs and individual functions, all in order to provide useful prompts for ChatGPT to act on. I learn best by trial and error, with the ability to ask why things happened or are the way they are.
So, as a secondary teaching assistant, I think it’s very useful. But trying to use the API for ChatGPT 4 is…not worth it. I can easily blow through $20 in a few hours. So, I got a day and a half of use out of it before I gave up. :|
ChatGPT is hilariously incompetent… but on a serious note, I still firmly reject tools like copilot outside demos and the like because they drastically reduce code quality for short term acceleration. That’s a terrible trade-off in terms of cost.
I enjoy using copilot, but it is not made to think for you. It’s a better autocomplete, but don’t ever let it do more than a line at once.
Yup, AI is a tool, not a complete solution.
As a software engineer, the number of people I encounter in a given week who either refuse to or are incapable of understanding that distinction baffles and concerns me.
That’s because it’s being advertised as a solution. That’s why you have people worried it’ll take their jobs when in reality it’ll let them do the job better.
The problem I have with it is that all the time it saves me I have to use on reading the code. I probably spend more time on that as once in a while the code it produces is broken in a subtle way.
I see some people swearing by it, which is the opposite of my experience. I suspect that if your coding was copying code from stack overflow then it indeed improved your experience as now this process is streamlined.
deleted by creator
Same as ChatGPT is better web search.
id rather search the web than chatgpt because i fact check it anyway
I don’t know if it does yet, but if ChatGPT starts providing source for every information, then it would make it much faster to find the relevant information and check their sources, rather than clicking websites one by one.
Yep, ChatGPT4 allows optional calls to Bing now.
It used to have a problem with making a claims that were not relevant to or contradicted its own sources, but I don’t recall encountering that problem recently.
Biggest problem with it is that it lies with the exact same confidence it tells the truth. Or, put another way, it’s confidently incorrect as often as it is confidently correct - and there’s no way to tell the difference unless you already know the answer.
it’s kinda hilarious to me because one of the FIRST things ai researchers did was get models to identify things and output answers together with the confidence of each potential ID, and now we’ve somehow regressed back from that point
did we really regress back from that?
i mean giving a confidence for recognizing a certain object in a picture is relatively straightforward.
But LLMs put together words by their likeliness of belonging together under your input (terribly oversimplified).the confidence behind that has no direct relation to how likely the statements made are true. I remember an example where someone made chatgpt say that 2+2 equals 5 because his wife said so. So chatgpt was confident that something is right when the wife says it, simply because it thinks these words to belong together.
ChatGPT what is the Gödel number for the proof of 2+2=5?
Of course I don’t know enough about the actual proof for it to be anything but a joke but there are infinite numbers so there should be infinite proofs.
there are also meme proofs out there I assume could be given a Gödel number easily enough.
Oh boy do I have news for you, that’s basically the only thing middle managers care about, short tem acceleration
But LinkedIn bros and corporate people are gonna gobble it up anyways because it has the right buzzwords (including “AI”) and they can squeeze more (low quality) work from devs to brag about how many things they (the corporate owners) are doing.
It’s just a fad. There’s just a small bit that will stay after the hype is gone. You know, like blockchain, AR, metaverse, NFT and whatever it was before that. In a few years there will be another breakthrough with it and we’ll hear from it again for a short while, but for now it’s just a one trick pony.
I always forget Metaverse is a thing.
deleted by creator
It was never a thing.
deleted by creator
I should’ve elaborated… It’s supposed to be a vast seamless network of interconnected spaces isn’t it, but that doesn’t exist yet. I think they jumped the gun in that everyone thought Horizons was it. It’s only just starting to be built.
I actually had ‘was’ first then edited after Googling. It appears it is still a thing but can’t really tell for what.
Is there really any utility for blockchain and NFTs?
The hilarious thing about blockchain is that the core concept is actively making the whole thing worse. The matrix protocol is sort of essentially blockchain without the decentralized ledger part, and it’s so vastly superior in every single way.
NFTs just show how fundamentally dumb blockchains are, if you skip the decentralized ledger bit then you never need to invent NFT functionality in the first place…
So a [Merkle tree](http://www…com/ https://en.wikipedia.org/wiki/Merkle_tree)?
i’m not sure whether matrix uses that, what i’m talking about is how it does the things everyone finds neat about blockchains without the inherent downsides like massive power usage and EVERYONE having to replicate the ENTIRE ledger.
i know IPFS uses merkle trees though, and hilariously blockchains largely rely on that to actually store any significantly sized data.
Well, if you stop listening to people who think it’s a way to get really rich really fast (which it obviously isn’t), cryptocurrencies are quite useful. International transfers are so much cheaper and easier with them.
Yes, you could, for example, use it to manage who is allowed to park in a garage, anonymously. The owner of a parking spot NFT can unlock the door from the outside. Stuff like that.
However, it’s also possible to do that with a small web application. Just payments and transfer of the parking spots are less free and it’s not decentralized.
The proton vpn people are either using or working on using block chain as a sort of email verification. Iirc it won’t have any cost or change in usage to the consumer, just an added layer of security. I’m not smart enough to understand how but it sounds neat.
It sounds like bullshit
I may have misread, Its not clear to me if it’ll be a staple of their free email service or will be added as part of their paid subscription to all their services but it still sounds neat. It very well could be bullshit but I’m interested in learning more.
https://www.tomsguide.com/news/proton-mail-to-use-blockchain-to-verify-recipients-email-addresses
I disagree, because unlike those things, ai actually has a use case. It needs a human supervisor and it isn’t always faster, but chat gpt has been the best educational resource I’ve ever had next to YouTube. It’s also decent at pumping out a lot of lazy work and writing so i don’t have to, or helping me break down a task into smaller parts. As long as you’re not expecting it to solve all your problems for you, it’s an amazing tool.
People said the same things about 3d printing and yeah, while it can’t create literally everything at industrial scale, and it’s not going to see much consumer use, it has found a place in certain applications like prototyping and small scale production.
ChatGPT is OK at summarizing popular, low specificity topics that tons of people have already written a ton about, but it’s terrible at anything else. When I tested its knowledge about the process of a niche interest of mine (fabric dyeing) it skipped completely over certain important pieces of information, and when I prompted it to include them it basically just mirrored my prompt back at me.
Which has pretty much summed up my ChatGPT experience: it just regurgitates stuff I can find myself, but removes the ability to determine if the source is reliable. And if it’s something I’m already having trouble finding detailed information about it usually doesn’t help.
i’d argue that chat gpt is mostly great at taking human bullshit tasks from humans, who dontwant to dothe bullshit, like regurgitating the text from a textbook in different words, writing cover letters for job applications, that are often machine analyzed for buzzwords anyways.
So its use case only exist in the domain of bullshit tasks that only exist to occupy two people without any added value.
Yeah, that is one thing it can do, but it’s not the only thing it does. I’m not sure how to get my point across well but, just because it gives you the wrong answer 25% of the time doesn’t mean it’s useless. In whatever you ask it to do, it often gets you most of the way there as long as you know how to correct it or check its work. The ability to ask specific or follow-up questions when learning something makes it invaluable as a learning tool (if you’re teaching yourself that is (ie. If you’re a university student)). It’s also very useful when brainstorming ideas or helping you approach a problem from a different angle. I can also ask it questions that are far more specific than what a search engine would get me.
It really comes down to if the human operator knows how (and when) to use it properly.
i never trust that information id rather learn with a book videos or just from websites if i ever use it i always fact check it. never blindly trust a computer
That was an integral part to this whole thing, you always fact check the ai. I said this in both of my comments.
I totally get preferring textbooks or videos though. I just find that the ai saves me time since i can ask specific questions about things, and it often gives me concise information that i understand more quickly.
Yeah, they think it can turn a beginner dev into an advanced dev, but really it’s more like having a team of beginner devs.
It’s alright for translation. As an intermediate dev, being able to translate knowledge into languages I’m not as familiar with is nice.
I’m still convinced that GitHub copilot is actively violating copyleft licenses. If not in word, then in the spirit.
Copyleft violences.
Western society is built on this principle
Tell me about it…
I left my more mature company for a startup.
I feel like Tyler Durden sometimes.
How you liking it? How many years have you aged in the months working at your startup?
My hairline has started receding very rapidly. There’s there’s these fine hairs all over my desk, and I see the photo I took when joining directly before turning on my camera every meeting.
Doesn’t sood good at all. I’m sorry to hear that, friend. I really hope there’s enough upsides there compared to working at a more mature company for you.
Sort of. Nobody’s cutting corners on aviation structural components, for example. We’ve been pretty good at maximizing general value output, and usually that means lower quality, but not always.
May want to show your roll a bit…
https://fortune.com/2023/10/03/delta-fourth-major-us-airline-fake-jet-aircraft-engine-parts-forged-airworthiness-documents-uk-company-aog/
I’m going to say that’s the exception that proves the rule, assuming they were structural parts and not a minor controller chip for de-icing or something.
The company themself announced it without being prompted, and if whoever introduce these unapproved parts into a small number of engines is caught there’s going to be real hell to pay. The stuff that stops you from falling out of the sky is serious business, and is largely treated as such.
On the other hand, a software function that’s hacked together and inefficient will just fly below the radar, and most people will prefer two cheap outfits to one that’s actually well made for the same price, so quality goes right out the window.
I’ve asked ChatGPT to fix simple bugs that were in the code it’s given me. It usually doesn’t understand that there’s a problem and just rewrites it with different variables names.
It’s helped me a bit with resolving weird tomcat/Java issues when upgrading to RHEL8, though. It didn’t give me an answer, but it gave me ideas on where to look (in my case I didn’t realize fapolicyd replaced selinux)
That’s the point - you have the expertise to make proper sense of whatever it outputs. The people pushing for “AI” the most want to rely on it without any necessary expertise or just minimal efforts, like feeding it some of your financial reports and have generate a 5-year strategy only to fail miserably and have no one to blame this time (will still blame anyone else but themselves btw).
It’s not the most useless tool in the world by any means, but the mainstream talk is completely out of touch with reality on the matter, and so are mainstream actions (i.e. overrelying on it and putting way too much faith into it).
Yeah, it can speed up the process but you still have to know how to do it yourself when it inevitably messes something up.
An unpopular opinion, I am sure, but if you’re a beginner with something - a new language, a new framework - and hate reading the docs, it’s a great way of just jumping into a new project. Like, I’ve been hacking away on a django web server for a personal project and it saved me a huge amount of time with understanding how apps are structured, how to interact with its settings, registering urls, creating views, the general development lifecycle of the project and the basic commands I need to do what I’m trying to do. God knows Google is a shitshow now and while Stackoverflow is fine and dandy (when it isn’t remarkably toxic and judgmental), the fact is that it cuts down on hours of fruitless research, assuming you’re not asking it to do anything genuinely novel or hyper-specific.
It helps a complete newbie like me get started and even learn while I do. Due to its restrictions and shortcoming, I’ve been having to learn how to structure and plan a project more carefully and thoughtfully, even creating design specs for programs and individual functions, all in order to provide useful prompts for ChatGPT to act on. I learn best by trial and error, with the ability to ask why things happened or are the way they are.
So, as a secondary teaching assistant, I think it’s very useful. But trying to use the API for ChatGPT 4 is…not worth it. I can easily blow through $20 in a few hours. So, I got a day and a half of use out of it before I gave up. :|