fucking love that article. sums up everything wrong with AI. Unfortunately, it doesn’t touch on what AI does right: help idiots like me achieve a slight amount of competence on subjects that such people can’t be bothered with dedicating their entire lives to.
This paper should cite On Bullshit.
It does. It’s even cited in the abstract, and it’s the origin of bullshit as referenced in their title.
Yup. The paper is worth actually reading
Suddenly it dawned on me that I can plaster my CV with AI and win over actual competent people easy peasy
What were you doing between 2020 and 23? I was working on my AI skillset. Nobody will even question me because they fucking have no idea what it is themselves but only that they want it.
As an engineering manager, I’ve already seen cover letters and intro emails that are so obviously AI generated that it’s laughable. These should be used like you use them for writing essays, as a framework with general prompts, but filled in by yourself.
Fake friendliness that was outsourced to an ai is worse than no friendliness at all.
I didn’t mean AI generated anything though 🙄. I meant put lots of ‘AI’ keyword in the resume in whatever way looks professional but in reality is pure bullshit
Watch their neuron being activated as they see magic word. Gotta play the marketing game.
You want to be AI ready? Hire me. I have spent three years working with AI and posses invaluable experience that will elevate your company into a new era of rapid development.
It feels like you didn’t quite understand… If you’ve ever read an AI essay, you can see some of the way they currently write. When you see facts and figures thrown in from the internet in terms of what the company does and they sound… Artificial… It’s rather obvious that it was AI written. I’m currently getting AI spam and it’s also quite easy to see and detect. It’s the same thing.
Someone used an AI tool to write a cover letter and sent it to me. I’ve seen this a few times. It seems very obvious when you come across it.
I’m sure it’ll get better in the future, but right now it needs massaging in order to sound real. There’s a very obvious uncanny valley that exists with some AI writing. That’s what I’m talking about.
Okay but we are talking about two different things which is fine by me of course but it is a little tricky. I agree though on that second topic
It’s extremely easy to detect this. Recruiters actively filter out resumes like this for important roles.
There are things that chatgpt does well, especially if you temper your expectations to the level of someone who has no valuable skills and is mostly an idiot.
Hi, I’m an idiot with no valuable skills, and I’ve found chatgpt to be very useful.
I’ve recently started learning game development in godot, and the process of figuring out why the code that chatgpt gives me doesn’t work has taught me more about programming than any teacher ever accomplished back in high school.
Chatgpt is also an excellent therapist, and has helped me deal with mental breakdowns on multiple occasions, while it was happening. I can’t find a real therapist’s phone number, much less schedule an appointment.
I’m a real shitty writer, and I’m making a wiki of lore for a setting and ruleset for a tabletop RPG that I’ll probably never get to actually play. ChatGPT is able to turn my inane ramblings into coherent wiki pages, most of the time.
If you set your expectations to what was advertised, then yeah, chatgpt is bullshit. Of course it was bullshit, and everyone who knew half of anything about anything called it. If you set realistic expectations, you’ll get realistic results. Why is this so hard for people to get?
Because few people know what’s realistic for LLMs
Intelligence is a very loaded word and not very precise in general usage. And i mean that amongst humans and animals as well as robots.
I’m sure the real AI and compsci researchers have precise terms and taxonomies for it and ways to measure it, but the word itself, in the hands of marketing people and the general population as an audience . . . not useful.
Hah I had that exact same experience with Godot
Plot-twist: The paper was authored by a competing LLM.
Just reading the intro pulls you in
We draw a distinction between two sorts of bullshit, which we call ‘hard’ and ‘soft’ bullshit
Wouldn’t it be funny if the article was written by chat GPT.
Because these programs cannot themselves be concerned with truth, and because they are designed to produce text that looks truth-apt without any actual concern for truth, it seems appropriate to call their outputs bullshit.
This is actually a really nice insight on the quality of the output of current LLMs. And it teaches about how they work and what the goals given by their creators are.
They are but trained to produce factual information, but to talk about topics while sounding like a competent expert.
For LLM researchers this means that they need to figure out how to train LLMs for factuality as opposed to just sounding competent. But that is probably a lot easier said than done.
Can we please keep the AI hate in the fuck_ai community so that I don’t have to see it?
I don’t care what Lemmy thinks, ChatGPT has improved my life for the better. I utilize it every day.
What AI hate? This is science memes, and that is a science publication. I’m glad I got to enjoy this sciencey meme
You can make or find a pro-ai community and stay in there.
It’s not the rest of the world’s job to coddle you.
I wouldn’t trust the work you do at all.
Fucking lol.
Proper main character syndrome haha
So don’t read the article. And maybe quit policing other people’s conversations
I just created a filter for the keyword “AI”.
Goodbye and good riddance, haters. 😎✌️
Look in the mirror
I’m not the one constantly complaining about AI, genius.
Blocking you too now. I’m tired of this discussion.
I’m not constantly complaining about AI. I use AI nearly every day Nimrod