The 200 year-old company may soon go public on the back of AI-powered education products.

  • PhilipTheBucket@ponder.cat
    link
    fedilink
    English
    arrow-up
    49
    arrow-down
    2
    ·
    3 days ago

    More general-purpose models like ChatGPT suffer from hallucinations because they have hoovered up the entire internet, including all the junk and misinformation.

    Incorrect. ChatGPT hallucinates because that’s how LLMs work. Hoovering up misinformation is a separate problem.

    A company in the space of selling educational books that has seen its fortunes go the opposite direction is Chegg. The company has seen its stock price plummet almost in lock-step with the rise of OpenAI’s ChatGPT, as students canceled their subscriptions to its online knowledge platform.

    Incorrect. Chegg is a cheating platform. It is the opposite of a knowledge platform.

    Why is Gizmodo paying people to write articles who apparently know pretty much nothing about the subject they are writing about?

      • PhilipTheBucket@ponder.cat
        link
        fedilink
        English
        arrow-up
        3
        ·
        3 days ago

        I read some of the author’s other articles. They have a habit of regurgitating highly suspect claims from press releases or company self-descriptions as if they were reality.

        OpenAI is confident in o3, and offers impressive benchmarks—it says that in a Codeforcing testing, which measures coding ability, o3 got a score of 2727. For context, a score of 2400 would put an engineer in the 99th percentile of programmers. It gets a score of 96.7% on the 2024 American Invitational Mathematics Exam, missing just one question.

        There’s also the article which claims that AI puts the entire power grid at risk, and then when you read the article, you learn that in order for that to be true, you need to lump AI in with crypto mining, other datacenter expansion, electric cars, and climate control for people’s homes.

  • Diplomjodler@lemmy.world
    link
    fedilink
    English
    arrow-up
    52
    ·
    edit-2
    2 days ago

    I miss the days when you could just do a thing and as long as you made more money than you spent, just keep doing it. Now the line must always go up, no matter what. And if your business model isn’t profitable enough they’ll just shut down 200 years of tradition without a second thought.

    • then_three_more@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      1
      ·
      3 days ago

      Surely companies have always adapted and changed products. They’re not exactly able to make much of encyclopedias anymore when they’re main competitor is free to use and has volunteers writing and editing it.

  • floofloof@lemmy.ca
    link
    fedilink
    English
    arrow-up
    37
    ·
    3 days ago

    Encyclopaedia Britannica—now known as just Britannica— is all in on artificial intelligence, and may soon go public at a valuation of nearly $1 billion

    All in on enshittification, like everyone else.

  • jagermo@feddit.org
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    17
    ·
    3 days ago

    I mean, this makes sense and could actually be a positive thing to look stuff up

    • Cuervo@lemmygrad.ml
      link
      fedilink
      English
      arrow-up
      4
      ·
      3 days ago

      considering the nature of an encyclopedia - short, knowledgeable articles on practically every topic - what’s the difference except to invite errors and snazz things up ?

    • FaceDeer@fedia.io
      link
      fedilink
      arrow-up
      4
      arrow-down
      7
      ·
      3 days ago

      How dare you say something insufficiently negative about the stuff everyone hates.

      • Lemminary@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        1
        ·
        2 days ago

        The downvotes are for the naïveté of the statement. Many people here use LLMs every day and have stated so in other threads. We just don’t think this is necessarily a proper use case given that you’re dealing with factual information. You can see as much in other comments on this thread pointing out the hallucinations.

        • FaceDeer@fedia.io
          link
          fedilink
          arrow-up
          3
          ·
          2 days ago

          Whereas I use LLMs every day, have actually written code that uses them, and I understand that they’re perfectly fine dealing with factual information when used in the proper framework. You’d be using retrieval-augmented generation (RAG) in an application like this.

          The “but hallucinations!” Objection goes in the same bin as “they can’t do fingers.” It’s an old concern that’s had a lot of work done to resolve it but that the general public haven’t bothered to keep up with.

          • Lemminary@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            2 days ago

            “they can’t do fingers.” It’s an old concern

            Have you seen those gorilla hands, though? Yes, there are five fingers there but everyone got fucking man hands. lmao

            It seems RAG helps mitigate but doesn’t eliminate hallucinations yet. Not to mention it’s quite expensive and has trouble extracting information based on abstract concepts. It sounds promising but it’s not the silver bullet I’m being sold on.