<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Blog on Hanyuan Jiang</title><link>https://jianghanyuan.github.io/blog/</link><description>Recent content in Blog on Hanyuan Jiang</description><generator>Hugo -- 0.156.0</generator><language>en-us</language><lastBuildDate>Mon, 02 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://jianghanyuan.github.io/blog/index.xml" rel="self" type="application/rss+xml"/><item><title>Humanity’s Thousand-Year Alignment Experiment</title><link>https://jianghanyuan.github.io/blog/humanitys-thousand-year-alignment-experiment/</link><pubDate>Mon, 02 Mar 2026 00:00:00 +0000</pubDate><guid>https://jianghanyuan.github.io/blog/humanitys-thousand-year-alignment-experiment/</guid><description>&lt;p&gt;When we discuss the problem of AI alignment, we tend to view it as an unprecedented technological challenge. However, human society has been conducting an alignment experiment for thousands of years. The object of this experiment is not silicon-based intelligence, but carbon-based intelligence itself. We call it law.&lt;/p&gt;
&lt;p&gt;In fact, I think the core dilemma of aligning legal systems with AI is strikingly similar. How to constrain the infinite possible behavior of an agent with finite rules? How to maintain the predictability of the system while pursuing justice? How do you balance the need for strict adherence to norms with the need for flexibility to respond to situations? More fundamentally, what does alignment mean when we can&amp;rsquo;t even agree on the &amp;ldquo;goal of alignment&amp;rdquo; itself? After a long evolution from Hammurabi&amp;rsquo;s Code to modern constitutional law, this experiment has not produced a perfect solution, but I think the lessons it has accumulated do provide a definitive answer to the problem of artificial intelligence alignment. It&amp;rsquo;s a never-ending quest.&lt;/p&gt;</description></item><item><title>The Rating You See Is Pricing Tomorrow</title><link>https://jianghanyuan.github.io/blog/the-rating-you-see-is-pricing-tomorrow/</link><pubDate>Thu, 26 Feb 2026 00:00:00 +0000</pubDate><guid>https://jianghanyuan.github.io/blog/the-rating-you-see-is-pricing-tomorrow/</guid><description>&lt;p&gt;In competitive games like &lt;a href="https://en.wikipedia.org/wiki/League_of_Legends"&gt;&lt;em&gt;League of Legends&lt;/em&gt;&lt;/a&gt; or &lt;a href="https://en.wikipedia.org/wiki/Valorant"&gt;&lt;em&gt;Valorant&lt;/em&gt;&lt;/a&gt;, your visible rank updates after every match. The standard story says that number measures skill. But something &amp;lsquo;stranger&amp;rsquo; happens the moment the same number also determines who you face next, which queues you can enter, and whether your friends can still play with you.&lt;/p&gt;
&lt;p&gt;Imagine a new ranked arena with clean outcomes: two players, one match, one rating that rises when you win and falls when you lose. The official explanation sounds familiar. The number tracks your underlying strength. This is the good old &lt;a href="https://en.wikipedia.org/wiki/Elo_rating_system"&gt;Elo&lt;/a&gt; dream in modern UI, and recent theory gives that dream real substance by showing that Elo can be understood as a serious online learning rule under a Bradley-Terry model.&lt;sup&gt;1&lt;/sup&gt;&lt;span class="sidenote"&gt;&lt;sup&gt;1&lt;/sup&gt; Olesker-Taylor and Zanetti (&lt;a href="#ref-olesker2024"&gt;2024&lt;/a&gt;) analyze Elo through Markov chains and formalize when it tracks latent skill rather than just acting as competitive folklore.&lt;/span&gt; Yet the moment the rating also allocates your next opponents and your next set of options, it begins to do more than estimate what you are.&lt;/p&gt;</description></item><item><title>Research Interests</title><link>https://jianghanyuan.github.io/blog/research-interests/</link><pubDate>Sun, 22 Feb 2026 00:00:00 +0000</pubDate><guid>https://jianghanyuan.github.io/blog/research-interests/</guid><description>&lt;p&gt;My research focuses on machine learning, especially frontier deep learning and reinforcement learning architecture, as well as LLM alignment. I am also interested in game theory and mechanism design.&lt;/p&gt;
&lt;p&gt;Drafting in progress&amp;hellip;&lt;/p&gt;</description></item><item><title>On Learning, Longing, and All the Ideas We Cannot Name</title><link>https://jianghanyuan.github.io/blog/on-learning-longing-and-all-the-ideas-we-cannot-name/</link><pubDate>Fri, 20 Feb 2026 00:00:00 +0000</pubDate><guid>https://jianghanyuan.github.io/blog/on-learning-longing-and-all-the-ideas-we-cannot-name/</guid><description>&lt;p&gt;There are nights when the world feels almost structured enough to reveal its secret. I lie awake thinking about the quiet impossibility at the center of learning. A child hears scattered fragments of language and somehow extracts the grammar of an entire tongue. A bird sees the stars rotating overhead and knows which direction to migrate. A mathematician stares at symbols until patterns crystallize that were always there but never visible. Structure appears where none was visibly given. Something in the mind finds what the world does not openly display.&lt;/p&gt;</description></item></channel></rss>