Stories
Slash Boxes
Comments

SoylentNews is people

posted by janrinok on Saturday April 08, @04:08PM   Printer-friendly

OpenAI threatened with landmark defamation lawsuit over ChatGPT false claims

https://arstechnica.com/tech-policy/2023/04/openai-may-be-sued-after-chatgpt-falsely-says-aussie-mayor-is-an-ex-con/

A spokesperson for Gordon Legal provided a statement to Ars confirming that responses to text prompts generated by ChatGPT 3.5 and 4 vary, with defamatory comments still currently being generated in ChatGPT 3.5. Among "several false statements" generated by ChatGPT were falsehoods stating that Brian Hood "was accused of bribing officials in Malaysia, Indonesia, and Vietnam between 1999 and 2005, that he was sentenced to 30 months in prison after pleading guilty to two counts of false accounting under the Corporations Act in 2012, and that he authorised payments to a Malaysian arms dealer acting as a middleman to secure a contract with the Malaysian Government." Because "all of these statements are false," Gordon Legal "filed a Concerns Notice to OpenAI" that detailed the inaccuracy and demanded a rectification. "As artificial intelligence becomes increasingly integrated into our society, the accuracy of the information provided by these services will come under close legal scrutiny," James Naughton, Hood's lawyer, said, noting that if a defamation claim is raised, it "will aim to remedy the harm caused" to Hood and "ensure the accuracy of this software in his case.")

It was only a matter of time before ChatGPT—an artificial intelligence tool that generates responses based on user text prompts—was threatened with its first defamation lawsuit. That happened last month, Reuters reported today, when an Australian regional mayor, Brian Hood, sent a letter on March 21 to the tool's developer, OpenAI, announcing his plan to sue the company for ChatGPT's alleged role in spreading false claims that he had gone to prison for bribery.

To avoid the landmark lawsuit, Hood gave OpenAI 28 days to modify ChatGPT's responses and stop the tool from spouting disinformation.

ChatGPT invented a sexual harassment scandal and named a real law prof as the accused

https://www.washingtonpost.com/technology/2023/04/05/chatgpt-lies/

Archive link: https://archive.is/lJj3c

One night last week, the law professor Jonathan Turley got a troubling email. As part of a research study, a fellow lawyer in California had asked the AI chatbot ChatGPT to generate a list of legal scholars who had sexually harassed someone. Turley's name was on the list.

The chatbot, created by OpenAI, said Turley had made sexually suggestive comments and attempted to touch a student while on a class trip to Alaska, citing a March 2018 article in The Washington Post as the source of the information. The problem: No such article existed. There had never been a class trip to Alaska. And Turley said he'd never been accused of harassing a student.

A regular commentator in the media, Turley had sometimes asked for corrections in news stories. But this time, there was no journalist or editor to call — and no way to correct the record.

"It was quite chilling," he said in an interview with The Post. "An allegation of this kind is incredibly harmful."

ChatGPT vs Google Bard: Which is better? We put them to the test.

https://arstechnica.com/information-technology/2023/04/clash-of-the-ai-titans-chatgpt-vs-bard-in-a-showdown-of-wits-and-wisdom/

In today's world of generative AI chatbots, we've witnessed the sudden rise of OpenAI's ChatGPT, introduced in November, followed by Bing Chat in February and Google's Bard in March. We decided to put these chatbots through their paces with an assortment of tasks to determine which one reigns supreme in the AI chatbot arena. Since Bing Chat uses similar GPT-4 technology as the latest ChatGPT model, we opted to focus on two titans of AI chatbot technology: OpenAI and Google.

We tested ChatGPT and Bard in seven critical categories: dad jokes, argument dialog, mathematical word problems, summarization, factual retrieval, creative writing, and coding. For each test, we fed the exact same instruction (called a "prompt") into ChatGPT (with GPT-4) and Google Bard. We used the first result, with no cherry-picking. Obviously, this is not a scientific study and is intended to be a fun comparison of the chatbots' capabilities. Outputs can vary between sessions due to random elements, and further evaluations with different prompts will produce different results. Also, the capabilities of these models will change rapidly over time as Google and OpenAI continue to upgrade them. But for now, this is how things stand in early April 2023.[....]


Original Submission #1Original Submission #2Original Submission #3Original Submission #4

 
This discussion was created by janrinok (52) for logged-in users only, but now has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
  • (Score: 5, Insightful) by stormreaver on Saturday April 08, @04:34PM (5 children)

    by stormreaver (5101) on Saturday April 08, @04:34PM (#1300516)

    Large Language Models are great at forming coherent-sounding sentences, but they absolutely suck at accuracy. With any luck, those accuracy screwups will eventually be so severe as to get those LLM's sued back into their natural niche: entertainment. And hopefully we will be able to go another 30 years without having the hear about "AI" nonsense again -- until the AI hype cycle is renewed amongst a new, young crowd who has failed to learn from history (like we're experiencing now).

    Starting Score:    1  point
    Moderation   +3  
       Insightful=3, Total=3
    Extra 'Insightful' Modifier   0  
    Karma-Bonus Modifier   +1  

    Total Score:   5  
  • (Score: 2) by takyon on Saturday April 08, @08:35PM (3 children)

    by takyon (881) <takyonNO@SPAMsoylentnews.org> on Saturday April 08, @08:35PM (#1300546) Journal

    It's easy to forget that "AI" can also do fun things like fire your radiologist or enable the creation of an unprecedented surveillance state. The money flowing into the ChatGPT sector will land everywhere and improve society in exciting ways.

    As for LLMs, exponential amounts of hardware could produce linear improvements just good enough to keep them afloat. Or the tech giants could take common sense approaches, like hiring mechanical turks to assemble a massive wikibase of approved knowledge. Bonus points if you can rip a human brain out, put it in a jar, and hook it up directly to the LLM supercomputer.

    --
    [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
    • (Score: 2) by stormreaver on Sunday April 09, @02:03AM (2 children)

      by stormreaver (5101) on Sunday April 09, @02:03AM (#1300590)

      It's easy to forget that "AI" can also do fun things like fire your radiologist....

      It would take a fool of monumental proportions to use "AI" in place of a trained radiologist. The latter can tell the difference between a tumor reading and a speck of dust on the lense. The former would probably not.

      • (Score: 2) by takyon on Sunday April 09, @02:52AM

        by takyon (881) <takyonNO@SPAMsoylentnews.org> on Sunday April 09, @02:52AM (#1300595) Journal

        It will end up being less people doing more work (with the assistance of friendly AI).

        --
        [SIG] 10/28/2017: Soylent Upgrade v14 [soylentnews.org]
      • (Score: 2) by PiMuNu on Sunday April 09, @11:17AM

        by PiMuNu (3823) on Sunday April 09, @11:17AM (#1300622)

        Actually, image recognition is quite a good use for these glorified fitting routines. This sort of numerical analysis tool has been used by scientists for many decades.

  • (Score: 4, Informative) by crm114 on Saturday April 08, @08:37PM

    by crm114 (8238) Subscriber Badge on Saturday April 08, @08:37PM (#1300549)

    Well, Wikipedia says Eliza was released in 1966, which would be 57 years. Sigh.