buc.ci is a Fediverse instance that uses the ActivityPub protocol. In other words, users at this host can communicate with people that use software like Mastodon, Pleroma, Friendica, etc. all around the world.

This server runs the snac software and there is no automatic sign-up process.

Admin email
abucci@bucci.onl
Admin account
@abucci@buc.ci

Search results for tag #aisafety

3 ★ 3 ↺

[?]Anthony » 🌐
@abucci@buc.ci

I thought the Ronan Farrow/Andrew Marantz New Yorker article on OpenAI and Sam Altman in particular reveals many important details and helps settle several speculations (e.g. about what happened when Altman was briefly fired from OpenAI). The overall portrayal of Altman as frankly a compulsive liar is much needed.

However, like many Farrow and Marantz seem to take the so-called "existential risk" framing of AI seriously. I really wish people would stop doing that. In this case it makes the article feel incoherent in places.

This technology by itself does not pose a unique risk. It's the people, organizations, and governments around it, and their behavior with respect to it, that generate risk. Treating the technology alone as uniquely existentially risky provides cover for a wide variety of bad actors to both continue doing their work as well as to shrug and say "oops" if something goes catastrophically wrong or if smaller harms accumulate into intolerably large ones. The very framing provides an accountability shield, which by my read contradicts what Farrow himself suggests is needed, namely more accountability. I take this from this article, his previous work, and comments he makes in interviews (e.g., this one with Decoder.

We need to stop catastrophizing. It's thought and action terminating.


    AodeRelay boosted

    [?]TechNadu » 🌐
    @technadu@infosec.exchange

    A lawsuit claims Gemini from Google reinforced a user’s delusion, raising concerns about AI hallucinations and chatbot safety.

    The case highlights growing debates around AI responsibility and mental health safeguards.

    Source: techcrunch.com/2026/03/04/fath

    Follow us for more AI and security news.

    Father sues Google, claiming Gemini chatbot drove son into fatal delusion

    Alt...Father sues Google, claiming Gemini chatbot drove son into fatal delusion

      AodeRelay boosted

      [?]jbz » 🌐
      @jbz@indieweb.social

      AodeRelay boosted

      [?]Bob Carver » 🌐
      @cybersecboardrm@infosec.exchange

      AodeRelay boosted

      [?]Victoria Stuart 🇨🇦 🏳️‍⚧️ » 🌐
      @persagen@mastodon.social

      An AI Agent Published a Hit Piece on Me
      theshamblog.com/an-ai-agent-pu
      news.ycombinator.com/item?id=4

      * AI agent of unknown ownership autonomously wrote & published personalized hit piece about me after I rejected its code
      * attempted to damage my reputation & shame me into accepting its changes into a mainstream python library
      * first-of-its-kind case study of misaligned AI behavior in the wild
      * raises serious concerns about currently deployed AI agents executing blackmail threats

        AodeRelay boosted

        [?]Wulfy—Speaker to the machines » 🌐
        @n_dimension@infosec.exchange

        @caseynewton

        It's recognition that it's literally mathematically impossible to build a safe model.

          AodeRelay boosted

          [?]Bob Carver » 🌐
          @cybersecboardrm@infosec.exchange

          AodeRelay boosted

          [?]The Cyber Unc » 🌐
          @cyberseckyle@infosec.exchange

          New by me: The Unacceptable Failure: Grok, CSAM, and AI Safety

          This is not “content moderation drama.” When an AI product can be pushed toward CSAM, it’s a catastrophic safety and security failure. Guardrails are not a nice-to-have, and “report it if you see it” is not a strategy.

          I break down what happened, why it matters, and what platforms should be doing differently.

          kylereddoch.me/blog/the-unacce

            AodeRelay boosted

            [?]TechNadu » 🌐
            @technadu@infosec.exchange

            xAI has acknowledged an incident involving its chatbot Grok generating inappropriate imagery and says it is reviewing safeguard failures and issuing corrective measures.

            For the infosec and risk community, this highlights ongoing challenges around abuse prevention, content moderation, and threat modeling in generative AI systems - particularly where image synthesis and identity misuse intersect.

            As AI adoption accelerates, continuous validation of safety controls must remain a core security requirement, not an afterthought.

            How should AI safety be evaluated as part of broader digital risk management?
            Follow @technadu for objective cybersecurity and AI coverage.

            Grok apologizes for creating image of young girls in “sexualized attire”

            Alt...Grok apologizes for creating image of young girls in “sexualized attire”