Slash Boxes

SoylentNews is people

posted by janrinok on Thursday May 25 2023, @10:41PM   Printer-friendly

Eric Schmidt wants to prevent potential abuse of AI:

Add Eric Schmidt to the list of tech luminaries concerned about the dangers of AI. The former Google chief tells guests at The Wall Street Journal's CEO Council Summit that AI represents an "existential risk" that could get many people "harmed or killed." He doesn't feel that threat is serious at the moment, but he sees a near future where AI could help find software security flaws or new biology types. It's important to ensure these systems aren't "misused by evil people," the veteran executive says.

Schmidt doesn't have a firm solution for regulating AI, but he believes there won't be an AI-specific regulator in the US. He participated in a National Security Commission on AI that reviewed the technology and published a 2021 report determining that the US wasn't ready for the tech's impact.

Schmidt doesn't have direct influence over AI. However, he joins a growing number of well-known moguls who have argued for a careful approach. Current Google CEO Sundar Pichai has cautioned that society needs to adapt to AI, while OpenAI leader Sam Altman has expressed concern that authoritarians might abuse these algorithms. In March, numerous industry leaders and researchers (including Elon Musk and Steve Wozniak) signed an open letter calling on companies to pause AI experiments for six months while they rethought the safety and ethical implications of their work.

Original Submission

This discussion was created by janrinok (52) for logged-in users only, but now has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
  • (Score: 1) by Mezion on Friday May 26 2023, @04:42AM

    by Mezion (18509) on Friday May 26 2023, @04:42AM (#1308253)

    But what ARE the dangers of AI? Except for deepfakes, nobody seems to want to answer that.

    Because it is FUD for an agenda. First thought on this article is that he (was part of what) made AI development the current "next big thing".
    This is the next stage, where the early birds attempt to block each other, and to block new competition from starting up.
    He doesn't want regulation of their AI products, he wants roadblocks created for anyone else that might attempt to get started in the same field.

    What we really need is methods preventing companies from making requirements for their "marketing features".
    The current amount of "required" information that is really only required for marketing/sale of supplied information is ridiculous.