<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Gwern.net — AI: Safety, Alignment &amp; Philosophy</title>
    <link>https://gwern.net</link>
    <description>Pages from gwern.net in the 'AI: Safety, Alignment &amp; Philosophy' section</description>
    <lastBuildDate>Sat, 07 Mar 2026 22:52:09 +0000</lastBuildDate>
    <generator>GwernRSSBuilder/1.0</generator>
    <atom:link href="https://gwern.net/feed.xml" rel="self" type="application/rss+xml"/>
    <item>
      <title>The Meta-LW Doomsday Argument, by Gwern · Gwern.net</title>
      <link>https://gwern.net/blog/2025/meta-doomsday</link>
      <guid isPermaLink="true">https://gwern.net/blog/2025/meta-doomsday</guid>
      <description>N/A</description>
      <author>Gwern</author>
      <pubDate>Tue, 29 Apr 2025 00:00:00 +0000</pubDate>
      <dc:date>2025-04-30T00:00:00Z</dc:date>
      <category>essay</category>
    </item>
    <item>
      <title>Area Man Angry AI Not Silver Bullet, by Gwern, Gemini-2.5-pro, GPT-4-o3 · Gwern.net</title>
      <link>https://gwern.net/blog/2025/area-man</link>
      <guid isPermaLink="true">https://gwern.net/blog/2025/area-man</guid>
      <description>&lt;em&gt;Onion&lt;/em&gt;-style fiction satire of complaints about automation or scaling not proceeding fast enough for some individual with unstated comparisons or timelines. (Partially an LLM style-imitation writing exercise.)</description>
      <author>Gwern, Gemini-2.5-pro, GPT-4-o3</author>
      <pubDate>Sat, 19 Apr 2025 00:00:00 +0000</pubDate>
      <dc:date>2025-04-19T00:00:00Z</dc:date>
      <category>essay</category>
      <category>ai/scaling/economics</category>
      <category>economics/automation</category>
      <category>fiction/humor</category>
      <category>reinforcement-learning/robot</category>
    </item>
    <item>
      <title>Copyright &amp; AI · Gwern.net</title>
      <link>https://gwern.net/ai-copyright</link>
      <guid isPermaLink="true">https://gwern.net/ai-copyright</guid>
      <description>The moral application of copyright and intellectual property to AI is the minimum protection necessary to promote the progress of science and the useful arts.</description>
      <author>Gwern</author>
      <pubDate>Sun, 26 Nov 2023 00:00:00 +0000</pubDate>
      <dc:date>2025-03-18T00:00:00Z</dc:date>
      <category>essay</category>
      <category>ai</category>
      <category>economics/copyright</category>
    </item>
    <item>
      <title>Is OpenAI OK?, by Gwern · Gwern.net</title>
      <link>https://gwern.net/blog/2024/openai-mojo</link>
      <guid isPermaLink="true">https://gwern.net/blog/2024/openai-mojo</guid>
      <description>Has OA lost its mojo and ability to push AI paradigms in the wake of its hypergrowth and the OA Altman coup? Is OA OK? How would we know &amp;amp; what would it look like if it was not?</description>
      <author>Gwern</author>
      <pubDate>Wed, 19 Jun 2024 00:00:00 +0000</pubDate>
      <dc:date>2024-12-04T00:00:00Z</dc:date>
      <category>essay</category>
    </item>
    <item>
      <title>‘Winning’ AI Arms Races: Then What?, by Gwern · Gwern.net</title>
      <link>https://gwern.net/blog/2024/winning-arms-races</link>
      <guid isPermaLink="true">https://gwern.net/blog/2024/winning-arms-races</guid>
      <description>Challenge to proponents of AI arms races and Manhattan Projects: state plainly, what is your long-term exit plan? What does ‘success’ look like? Or what good is it to beat rivals to AGI and then stand by while they catch up?</description>
      <author>Gwern</author>
      <pubDate>Sat, 23 Nov 2024 00:00:00 +0000</pubDate>
      <dc:date>2024-11-24T00:00:00Z</dc:date>
      <category>essay</category>
    </item>
    <item>
      <title>The Existential Risk of Math Errors · Gwern.net</title>
      <link>https://gwern.net/math-error</link>
      <guid isPermaLink="true">https://gwern.net/math-error</guid>
      <description>Mathematical mistake/error-rates limit our understanding of rare risks and ability to defend against them.</description>
      <author>Gwern</author>
      <pubDate>Fri, 20 Jul 2012 00:00:00 +0000</pubDate>
      <dc:date>2024-10-20T00:00:00Z</dc:date>
      <category>essay</category>
      <category>ai</category>
      <category>insight-porn</category>
      <category>philosophy</category>
      <category>statistics/bias</category>
      <category>survey</category>
      <category>transhumanism</category>
    </item>
    <item>
      <title>What Is an ‘AI Warning Shot’?, by Gwern · Gwern.net</title>
      <link>https://gwern.net/blog/2024/sydney</link>
      <guid isPermaLink="true">https://gwern.net/blog/2024/sydney</guid>
      <description>N/A</description>
      <author>Gwern</author>
      <pubDate>Wed, 07 Aug 2024 00:00:00 +0000</pubDate>
      <dc:date>2024-08-26T00:00:00Z</dc:date>
      <category>essay</category>
    </item>
    <item>
      <title>It Looks Like You’re Trying To Take Over The World · Gwern.net</title>
      <link>https://gwern.net/fiction/clippy</link>
      <guid isPermaLink="true">https://gwern.net/fiction/clippy</guid>
      <description>Fictional short story about Clippy &amp;amp; AI hard takeoff scenarios grounded in contemporary ML scaling, self-supervised learning, reinforcement learning, and meta-learning research literature.</description>
      <author>Gwern</author>
      <pubDate>Sun, 06 Mar 2022 00:00:00 +0000</pubDate>
      <dc:date>2023-03-28T00:00:00Z</dc:date>
      <category>essay</category>
      <category>ai/nn/transformer/gpt/inner-monologue</category>
      <category>fiction/humor</category>
      <category>fiction/science-fiction</category>
      <category>reinforcement-learning/safe</category>
      <category>reinforcement-learning/scaling</category>
    </item>
    <item>
      <title>Evolution as Backstop for Reinforcement Learning · Gwern.net</title>
      <link>https://gwern.net/backstop</link>
      <guid isPermaLink="true">https://gwern.net/backstop</guid>
      <description>Markets/evolution as backstops/ground truths for reinforcement learning/optimization: on some connections between Coase’s theory of the firm/linear optimization/DRL/evolution/multicellular life/pain/Internet communities as multi-level optimization problems.</description>
      <author>Gwern</author>
      <pubDate>Thu, 06 Dec 2018 00:00:00 +0000</pubDate>
      <dc:date>2021-07-04T00:00:00Z</dc:date>
      <category>essay</category>
      <category>ai/nn</category>
      <category>economics/automation</category>
      <category>insight-porn</category>
      <category>philosophy/epistemology</category>
      <category>psychology/energy</category>
      <category>psychology/willpower</category>
      <category>reinforcement-learning/multi-agent</category>
      <category>reinforcement-learning/safe</category>
      <category>statistics/bayes</category>
      <category>statistics/decision</category>
      <category>technology</category>
    </item>
    <item>
      <title>Complexity no Bar to AI · Gwern.net</title>
      <link>https://gwern.net/complexity</link>
      <guid isPermaLink="true">https://gwern.net/complexity</guid>
      <description>Critics of AI risk suggest diminishing returns to computing (formalized asymptotically) means AI will be weak; this argument relies on a large number of questionable premises and ignoring additional resources, constant factors, and nonlinear returns to small intelligence advantages, and is highly unlikely.</description>
      <author>Gwern</author>
      <pubDate>Sun, 01 Jun 2014 00:00:00 +0000</pubDate>
      <dc:date>2019-06-09T00:00:00Z</dc:date>
      <category>essay</category>
      <category>ai</category>
      <category>economics/automation</category>
      <category>reinforcement-learning/safe</category>
      <category>statistics/prediction</category>
      <category>transhumanism</category>
    </item>
    <item>
      <title>The Hyperbolic Time Chamber &amp; Brain Emulation · Gwern.net</title>
      <link>https://gwern.net/hyperbolic-time-chamber</link>
      <guid isPermaLink="true">https://gwern.net/hyperbolic-time-chamber</guid>
      <description>A time dilation chamber as thought experiment on the power of pure thought, with comparison to computer AGI advantages/disadvantages.</description>
      <author>Gwern</author>
      <pubDate>Wed, 29 Aug 2012 00:00:00 +0000</pubDate>
      <dc:date>2018-09-02T00:00:00Z</dc:date>
      <category>essay</category>
      <category>ai/scaling/economics</category>
      <category>anime</category>
      <category>economics/automation</category>
      <category>fiction/science-fiction</category>
      <category>transhumanism</category>
    </item>
    <item>
      <title>Why Tool AIs Want to Be Agent AIs · Gwern.net</title>
      <link>https://gwern.net/tool-ai</link>
      <guid isPermaLink="true">https://gwern.net/tool-ai</guid>
      <description>AIs limited to pure computation (Tool AIs) supporting humans, will be less intelligent, efficient, and economically valuable than more autonomous reinforcement-learning AIs (Agent AIs) who act on their own and meta-learn, because all problems are reinforcement-learning problems.</description>
      <author>Gwern</author>
      <pubDate>Wed, 07 Sep 2016 00:00:00 +0000</pubDate>
      <dc:date>2018-08-28T00:00:00Z</dc:date>
      <category>essay</category>
      <category>ai/scaling/economics</category>
      <category>economics/automation</category>
      <category>existential-risk</category>
      <category>insight-porn</category>
      <category>reinforcement-learning/safe</category>
      <category>reinforcement-learning/scaling</category>
    </item>
    <item>
      <title>Simulation Inferences · Gwern.net</title>
      <link>https://gwern.net/simulation-inference</link>
      <guid isPermaLink="true">https://gwern.net/simulation-inference</guid>
      <description>How small must be the computer simulating the universe?</description>
      <author>Gwern</author>
      <pubDate>Fri, 29 May 2009 00:00:00 +0000</pubDate>
      <dc:date>2012-04-15T00:00:00Z</dc:date>
      <category>essay</category>
      <category>cs</category>
      <category>math</category>
      <category>philosophy</category>
      <category>transhumanism</category>
    </item>
  </channel>
</rss>
