what are you whining about? Hallucination is inherently part of LLM as of today. Anything out of this should not be trusted with certainty. But employing it will have more benefits than just shadowing it for everyone. Take it as an unfinished project, so ignore the results if you like. Seriously, it’s physically possible to actually ignore the generative results.
I absolutely agree and I consider LLM results to be “neat” but never trusted - if I think I should bake spaghetti squash at 350 I might ask an LLM and only find real advice if our suggested temperatures vary.
But some people have wholly bought into the “it’s a magic knowledge box” bullshit - you’ll see opinions here on lemmy that generative AI can make novel creations that indicate true creativity… you’ll see opinions from C-level folks that LLMs can replace CS wholesale who are chomping at the bit to downsize call centers. Companies need to be careful about deceiving these users and those that feed into the mysticism really need to be stopped.
you’ll see opinions here on lemmy that generative AI can make novel creations that indicate true creativity…
Yeah I’m not jumping that bandwagon yet, but I think no one is able to determine either side of that. It makes terrific art, so it’s not out the realm of impossibility. The only sensible stance we can take right now, is none, and just wait and see whether AI art can hold up in the long run.
Taking any serious stance right now a priori would be illogical. I don’t understand the fuss people make it out to be. Yes, artists will suffer financially and will therefore limit their time investment and advancements in art, but sacrificing or halting the development of AI for them is also not a possibility. So yes, artists are fucked right now, but there is nothing we can do about that right now. Hopefully, some UBI, but that’s not here now.
But yes, companies deceiving and not warning about the hallucinations of AI, is bad. But it’s not their fault people believe stupid shit because they have always believed and will always believe stupid shit.
what are you whining about? Hallucination is inherently part of LLM as of today. Anything out of this should not be trusted with certainty. But employing it will have more benefits than just shadowing it for everyone. Take it as an unfinished project, so ignore the results if you like. Seriously, it’s physically possible to actually ignore the generative results.
“Absolutely sued” my ass
I absolutely agree and I consider LLM results to be “neat” but never trusted - if I think I should bake spaghetti squash at 350 I might ask an LLM and only find real advice if our suggested temperatures vary.
But some people have wholly bought into the “it’s a magic knowledge box” bullshit - you’ll see opinions here on lemmy that generative AI can make novel creations that indicate true creativity… you’ll see opinions from C-level folks that LLMs can replace CS wholesale who are chomping at the bit to downsize call centers. Companies need to be careful about deceiving these users and those that feed into the mysticism really need to be stopped.
Yeah I’m not jumping that bandwagon yet, but I think no one is able to determine either side of that. It makes terrific art, so it’s not out the realm of impossibility. The only sensible stance we can take right now, is none, and just wait and see whether AI art can hold up in the long run.
Taking any serious stance right now a priori would be illogical. I don’t understand the fuss people make it out to be. Yes, artists will suffer financially and will therefore limit their time investment and advancements in art, but sacrificing or halting the development of AI for them is also not a possibility. So yes, artists are fucked right now, but there is nothing we can do about that right now. Hopefully, some UBI, but that’s not here now.
But yes, companies deceiving and not warning about the hallucinations of AI, is bad. But it’s not their fault people believe stupid shit because they have always believed and will always believe stupid shit.