<?xml version="1.0" encoding="utf-8" standalone="yes" ?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Big AI Dream | Jie Fu</title>
    <link>/</link>
      <atom:link href="/index.xml" rel="self" type="application/rss+xml" />
    <description>Big AI Dream | Jie Fu</description>
    <generator>Hugo Blox Builder (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Mon, 24 Oct 2022 00:00:00 +0000</lastBuildDate>
    <image>
      <url>/media/icon_hu_e6925d744801042f.png</url>
      <title>Big AI Dream | Jie Fu</title>
      <link>/</link>
    </image>
    
    <item>
      <title></title>
      <link>/group/</link>
      <pubDate>Sat, 03 Jan 2026 00:00:00 +0000</pubDate>
      <guid>/group/</guid>
      <description></description>
    </item>
    
    <item>
      <title>A Definition of AGI</title>
      <link>/publication/agi/</link>
      <pubDate>Wed, 01 Oct 2025 00:00:00 +0000</pubDate>
      <guid>/publication/agi/</guid>
      <description>&lt;!-- 




  
    
  

&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;Create your slides in Markdown - click the &lt;em&gt;Slides&lt;/em&gt; button to check out the example.&lt;/span&gt;
&lt;/div&gt; --&gt;
&lt;!-- Supplementary notes can be added here, including [code, math, and images](https://wowchemy.com/docs/writing-markdown-latex/). --&gt;
</description>
    </item>
    
    <item>
      <title>Re:Form -- Reducing Human Priors in Scalable Formal Software Verification with RL in LLMs: A Preliminary Study on Dafny</title>
      <link>/publication/reform/</link>
      <pubDate>Wed, 23 Jul 2025 00:00:00 +0000</pubDate>
      <guid>/publication/reform/</guid>
      <description>&lt;!-- 




  
    
  

&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;Create your slides in Markdown - click the &lt;em&gt;Slides&lt;/em&gt; button to check out the example.&lt;/span&gt;
&lt;/div&gt; --&gt;
&lt;!-- Supplementary notes can be added here, including [code, math, and images](https://wowchemy.com/docs/writing-markdown-latex/). --&gt;
</description>
    </item>
    
    <item>
      <title>Autoformalization and Formally Verifiable AI</title>
      <link>/project/auto/</link>
      <pubDate>Sun, 25 May 2025 00:00:00 +0000</pubDate>
      <guid>/project/auto/</guid>
      <description>&lt;h2 id=&#34;motivation--goal&#34;&gt;Motivation &amp;amp; Goal&lt;/h2&gt;
&lt;p&gt;Autoformalization (converting natural language content into verifiable formalization) is considered a promising way of learning general purpose reasoning&lt;sup id=&#34;fnref:1&#34;&gt;&lt;a href=&#34;#fn:1&#34; class=&#34;footnote-ref&#34; role=&#34;doc-noteref&#34;&gt;1&lt;/a&gt;&lt;/sup&gt;.
In sharp contrast, existing natural language-based LLMs lack reliable verification, e.g., there seems to be no free lunch for inference scaling&lt;sup id=&#34;fnref:2&#34;&gt;&lt;a href=&#34;#fn:2&#34; class=&#34;footnote-ref&#34; role=&#34;doc-noteref&#34;&gt;2&lt;/a&gt;&lt;/sup&gt;, which is believed to be able to boost LLMs&amp;rsquo; reasoning capabilities, as the verifier (reliable training signals) is not perfect for most cases.
Formal verifiers are not only important for increasing the resiliency of humanity&lt;sup id=&#34;fnref:3&#34;&gt;&lt;a href=&#34;#fn:3&#34; class=&#34;footnote-ref&#34; role=&#34;doc-noteref&#34;&gt;3&lt;/a&gt;&lt;/sup&gt;, but also vital for steering the AI development into a direction I call &lt;code&gt;maximally math-seeking&lt;/code&gt;, which could be, hopefully, more human-friendly and realistic.&lt;/p&gt;
&lt;p&gt;Formal verification is usually extremely hard to obtain, and recent progress in automated reasoning could potentially make this approach easier.
However, existing LLMs cannot do genuine logical reasoning or self-verification on their own, and they should be viewed as universal approximate knowledge retrievers (e.g., trying to mimic the reasoning steps seen during training)&lt;sup id=&#34;fnref:4&#34;&gt;&lt;a href=&#34;#fn:4&#34; class=&#34;footnote-ref&#34; role=&#34;doc-noteref&#34;&gt;4&lt;/a&gt;&lt;/sup&gt;.
Given the important role of formal verifiers, I&amp;rsquo;m exploring ways of scaling up them.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;A more detailed blog post is &lt;a href=&#34;https://bigaidream.github.io/garden/autoformalization-and-formally-verifiable-AI&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;here&lt;/a&gt;&lt;/p&gt;&lt;/blockquote&gt;
&lt;!-- Following the formulation in [Distilling System 2 into System 1](https://arxiv.org/abs/2407.06023), given an input $x$, System-1 produces the output $y$ directly: $S_{\mathrm{I}}(x)=f_{\theta}(x)\to y\$. In contrast, System-2, takes an LLM $f_{\theta}$ and input $x$ and generates intermedaite tokens $z$: $S_\text{II}{(x;f_\theta)}\to z,y$, which can be seen as a form of meta learning. I plan to design scalable System-2 LLMs from the meta learning perspective.  --&gt;
&lt;!-- ### My (Remotely) Related Works

As of 2024-12, to be honest, I don&#39;t have works strongly related to what I&#39;m working on. 

#### Neural Architectures

- [Unlocking Emergent Modularity in Large Language Models](https://aclanthology.org/2024.naacl-long.144/), NAACL 2024 Outstanding Paper
- [Beyond Fully-Connected Layers with Quaternions: Parameterization of Hypercomplex Multiplications with $1/n$ Parameters](https://arxiv.org/abs/2102.08597), ICLR 2021 Outstanding Paper

#### Reinforcement Learning

- [Think Before You Act: Decision Transformers with Internal Working Memory](https://arxiv.org/abs/2305.16338), ICML 2024

#### Human-AI Alignment
- [AI Alignment: A Comprehensive Survey](https://arxiv.org/abs/2310.19852), arXiv 2023
- [Would you Rather? A New Benchmark for Learning Machine Alignment with Cultural Values and Social Preferences](https://aclanthology.org/2020.acl-main.477/), ACL 2020

#### Robustness
- [A survey of backdoor attacks and defenses on large language models: Implications for security measures](https://arxiv.org/abs/2406.06852), TMLR 2025, ([Survey Certification](https://jmlr.org/tmlr/papers/)) 
- [Prompt as Triggers for Backdoor Attack: Examining the Vulnerability in Language ModelsL](https://arxiv.org/abs/2305.01219), EMNLP 2023
- [Jacobian Adversarially Regularized Networks for Robustness](https://arxiv.org/abs/1912.10185), ICLR 2020


#### Meta Learning

- [Massive Editing for Large Language Models via Meta Learning](https://arxiv.org/abs/2311.04661), ICLR 2024
- [Learning Multi-Objective Curricula for Robotic Policy Learning](https://openreview.net/forum?id=ZL2keFk7WXJ), CoRL 2023
- [DrMAD: Distilling Reverse-Mode Automatic Differentiation for Optimizing Hyperparameters of Deep Neural Networks](https://arxiv.org/abs/1601.00917), IJCAI 2016 --&gt;
&lt;h2 id=&#34;approaches&#34;&gt;Approaches&lt;/h2&gt;
&lt;p&gt;We are focusing on the following approaches (not necessarily exclusive):&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Deep reinforcement learning&lt;/li&gt;
&lt;li&gt;Generative model, e.g., diffusion LM, autoregressive LM&lt;/li&gt;
&lt;li&gt;Formal verification&lt;/li&gt;
&lt;/ul&gt;
&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;We focus on scaling things up!&lt;/span&gt;
&lt;/div&gt;
&lt;h2 id=&#34;seeking-collaboration&#34;&gt;Seeking Collaboration&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;If you are a full-time researcher (familiar with formal verification, software engineering, or deep RL) and is interested in collaborating, I&amp;rsquo;d be very happy to chat.&lt;/li&gt;
&lt;li&gt;If you are a student (eager to learn formal verification, software engineering, and deep RL &amp;ndash; better already familiar with one of these), I could host you as an intern. See &lt;a href=&#34;/group/&#34;&gt;this page&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;credits&#34;&gt;Credits&lt;/h2&gt;
&lt;p&gt;Since this is a big and ambitious project, I believe that it is important to assign the correct amount credits to each member. I carefully divide it into a series of mini-projects. Each member is encouraged to take a lead in one of them, and each mini-project would hopefully result in an independent paper; More importantly, these mini-projects will be merged as the project to solve a big problem.&lt;/p&gt;
&lt;div class=&#34;footnotes&#34; role=&#34;doc-endnotes&#34;&gt;
&lt;hr&gt;
&lt;ol&gt;
&lt;li id=&#34;fn:1&#34;&gt;
&lt;p&gt;&lt;a href=&#34;https://research.google/pubs/a-promising-path-towards-autoformalization-and-general-artificial-intelligence/&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;https://research.google/pubs/a-promising-path-towards-autoformalization-and-general-artificial-intelligence/&lt;/a&gt;&amp;#160;&lt;a href=&#34;#fnref:1&#34; class=&#34;footnote-backref&#34; role=&#34;doc-backlink&#34;&gt;&amp;#x21a9;&amp;#xfe0e;&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&#34;fn:2&#34;&gt;
&lt;p&gt;&lt;a href=&#34;https://arxiv.org/abs/2411.17501&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;https://arxiv.org/abs/2411.17501&lt;/a&gt;&amp;#160;&lt;a href=&#34;#fnref:2&#34; class=&#34;footnote-backref&#34; role=&#34;doc-backlink&#34;&gt;&amp;#x21a9;&amp;#xfe0e;&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&#34;fn:3&#34;&gt;
&lt;p&gt;The plan described by &lt;a href=&#34;https://www.aria.org.uk/opportunity-spaces/mathematics-for-safe-ai/safeguarded-ai/&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;ARIA&lt;/a&gt; could potentially greatly increase the safety level but seems a bit over-constraining, which might introduce cancerous states as discussed in &lt;a href=&#34;https://arxiv.org/abs/2405.02325&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;this paper&lt;/a&gt;.&amp;#160;&lt;a href=&#34;#fnref:3&#34; class=&#34;footnote-backref&#34; role=&#34;doc-backlink&#34;&gt;&amp;#x21a9;&amp;#xfe0e;&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li id=&#34;fn:4&#34;&gt;
&lt;p&gt;&lt;a href=&#34;https://x.com/rao2z/status/1740692722099630237&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;https://x.com/rao2z/status/1740692722099630237&lt;/a&gt;&amp;#160;&lt;a href=&#34;#fnref:4&#34; class=&#34;footnote-backref&#34; role=&#34;doc-backlink&#34;&gt;&amp;#x21a9;&amp;#xfe0e;&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;/div&gt;
</description>
    </item>
    
    <item>
      <title>Finite State Automata Inside Transformers with Chain-of-Thought: A Mechanistic Study on State Tracking</title>
      <link>/publication/state_track/</link>
      <pubDate>Thu, 27 Feb 2025 00:00:00 +0000</pubDate>
      <guid>/publication/state_track/</guid>
      <description>&lt;!-- 




  
    
  

&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;Create your slides in Markdown - click the &lt;em&gt;Slides&lt;/em&gt; button to check out the example.&lt;/span&gt;
&lt;/div&gt; --&gt;
&lt;!-- Supplementary notes can be added here, including [code, math, and images](https://wowchemy.com/docs/writing-markdown-latex/). --&gt;
</description>
    </item>
    
    <item>
      <title>Unlocking Emergent Modularity in Large Language Models</title>
      <link>/publication/emoe/</link>
      <pubDate>Sat, 29 Jun 2024 02:00:00 +0000</pubDate>
      <guid>/publication/emoe/</guid>
      <description>&lt;!-- 




  
    
  

&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;Create your slides in Markdown - click the &lt;em&gt;Slides&lt;/em&gt; button to check out the example.&lt;/span&gt;
&lt;/div&gt; --&gt;
&lt;!-- Supplementary notes can be added here, including [code, math, and images](https://wowchemy.com/docs/writing-markdown-latex/). --&gt;
</description>
    </item>
    
    <item>
      <title>Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training</title>
      <link>/publication/stacking/</link>
      <pubDate>Thu, 30 May 2024 00:00:00 +0000</pubDate>
      <guid>/publication/stacking/</guid>
      <description>&lt;!-- 




  
    
  

&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;Create your slides in Markdown - click the &lt;em&gt;Slides&lt;/em&gt; button to check out the example.&lt;/span&gt;
&lt;/div&gt; --&gt;
&lt;!-- Supplementary notes can be added here, including [code, math, and images](https://wowchemy.com/docs/writing-markdown-latex/). --&gt;
</description>
    </item>
    
    <item>
      <title>Tracking single cell evolution via clock-like chromatin accessibility</title>
      <link>/publication/track_cell_evo/</link>
      <pubDate>Mon, 20 May 2024 02:00:00 +0000</pubDate>
      <guid>/publication/track_cell_evo/</guid>
      <description>&lt;!-- 




  
    
  

&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;Create your slides in Markdown - click the &lt;em&gt;Slides&lt;/em&gt; button to check out the example.&lt;/span&gt;
&lt;/div&gt; --&gt;
&lt;!-- Supplementary notes can be added here, including [code, math, and images](https://wowchemy.com/docs/writing-markdown-latex/). --&gt;
</description>
    </item>
    
    <item>
      <title>Think Before You Act: Decision Transformers with Working Memory</title>
      <link>/publication/think_act/</link>
      <pubDate>Sun, 19 May 2024 02:00:00 +0000</pubDate>
      <guid>/publication/think_act/</guid>
      <description>&lt;!-- 




  
    
  

&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;Create your slides in Markdown - click the &lt;em&gt;Slides&lt;/em&gt; button to check out the example.&lt;/span&gt;
&lt;/div&gt; --&gt;
&lt;!-- Supplementary notes can be added here, including [code, math, and images](https://wowchemy.com/docs/writing-markdown-latex/). --&gt;
</description>
    </item>
    
    <item>
      <title>Massive Editing for Large Language Models via Meta Learning</title>
      <link>/publication/malmen/</link>
      <pubDate>Tue, 16 Jan 2024 02:00:00 +0000</pubDate>
      <guid>/publication/malmen/</guid>
      <description>&lt;!-- 




  
    
  

&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;Create your slides in Markdown - click the &lt;em&gt;Slides&lt;/em&gt; button to check out the example.&lt;/span&gt;
&lt;/div&gt; --&gt;
&lt;!-- Supplementary notes can be added here, including [code, math, and images](https://wowchemy.com/docs/writing-markdown-latex/). --&gt;
</description>
    </item>
    
    <item>
      <title></title>
      <link>/contact/</link>
      <pubDate>Tue, 24 Oct 2023 00:00:00 +0000</pubDate>
      <guid>/contact/</guid>
      <description></description>
    </item>
    
    <item>
      <title></title>
      <link>/fun/</link>
      <pubDate>Tue, 24 Oct 2023 00:00:00 +0000</pubDate>
      <guid>/fun/</guid>
      <description></description>
    </item>
    
    <item>
      <title>Unifying Likelihood-free Inference with Black-box Sequence Design and Beyond</title>
      <link>/publication/unifying_likelihood/</link>
      <pubDate>Thu, 20 Jan 2022 00:00:00 +0000</pubDate>
      <guid>/publication/unifying_likelihood/</guid>
      <description>&lt;!-- 




  
    
  

&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;Create your slides in Markdown - click the &lt;em&gt;Slides&lt;/em&gt; button to check out the example.&lt;/span&gt;
&lt;/div&gt; --&gt;
&lt;!-- Supplementary notes can be added here, including [code, math, and images](https://wowchemy.com/docs/writing-markdown-latex/). --&gt;
</description>
    </item>
    
    <item>
      <title>Beyond Fully-Connected Layers with Quaternions: Parameterization of Hypercomplex Multiplications with 1/n Parameters</title>
      <link>/publication/parameterization_hypercomplex/</link>
      <pubDate>Fri, 02 Apr 2021 00:00:00 +0000</pubDate>
      <guid>/publication/parameterization_hypercomplex/</guid>
      <description>&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;Our paper received the &lt;a href=&#34;https://iclr-conf.medium.com/announcing-iclr-2021-outstanding-paper-awards-9ae0514734ab&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;Outstanding Paper Award&lt;/a&gt; (8 out of 860 accepted papers).&lt;/span&gt;
&lt;/div&gt;
&lt;!-- 




  
    
  

&lt;div class=&#34;flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900&#34;&gt;
&lt;span class=&#34;pr-3 pt-1 text-primary-600 dark:text-primary-300&#34;&gt;
  &lt;svg height=&#34;24&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z&#34;/&gt;&lt;/svg&gt;
&lt;/span&gt;
  &lt;span class=&#34;dark:text-neutral-300&#34;&gt;Create your slides in Markdown - click the &lt;em&gt;Slides&lt;/em&gt; button to check out the example.&lt;/span&gt;
&lt;/div&gt; --&gt;
&lt;!-- Supplementary notes can be added here, including [code, math, and images](https://wowchemy.com/docs/writing-markdown-latex/). --&gt;
</description>
    </item>
    
    <item>
      <title>Mentoring Junior Students</title>
      <link>/page/mentor/</link>
      <pubDate>Sun, 08 Nov 2020 00:00:00 +0000</pubDate>
      <guid>/page/mentor/</guid>
      <description>&lt;p&gt;I &amp;#x1f639; am always passionate about talking / mentoring / working with junior self-motivated students. I prefer writing detailed documents so that team members can discuss the technical details more easily.&lt;/p&gt;
&lt;!-- I&#39;m a hands-on senior postdoc: I&#39;d like to give very concrete suggestions and have frequent (can be short) 1-1 meetings with junior student.  --&gt;
&lt;p&gt;Drop me an email if you want to have a chat.&lt;/p&gt;
&lt;p&gt;I feel honored to be able to work, and grow together, with many young students with unlimited talent.
Here is a list of students that I am co-mentoring or co-mentored with other senior researchers:&lt;/p&gt;
&lt;h1 id=&#34;current&#34;&gt;Current&lt;/h1&gt;
&lt;ul&gt;
&lt;li&gt;Zixuan Liu, Drug Discovery, PhD Student at University of Washington&lt;/li&gt;
&lt;li&gt;Dan Liu, Meta Learning, PhD Student at McGill University&lt;/li&gt;
&lt;li&gt;Jikun Kang, Meta RL , PhD Student at McGill University&lt;/li&gt;
&lt;li&gt;Shaoxiong Ji, Protein LM, PhD Student at Aalto University&lt;/li&gt;
&lt;li&gt;Osana Ratnaharan, Health, Undergraduate Student at University of Toronto&lt;/li&gt;
&lt;li&gt;Qifeng Wu, Drug Discovery, Master Student at Fudan University&lt;/li&gt;
&lt;li&gt;Maolong Yang, Meta Learning, Undergrad Student at Tsinghua University&lt;/li&gt;
&lt;li&gt;Zedian Xiao, Drug Discovery, Undergrad Student at McGill University&lt;/li&gt;
&lt;li&gt;Haodong Ling, Drug Discovery, Undergrad Student at Fudan University&lt;/li&gt;
&lt;/ul&gt;
&lt;h1 id=&#34;2020&#34;&gt;2020&lt;/h1&gt;
&lt;ul&gt;
&lt;li&gt;Bingchan Zhao, RL for power network, undergraduate at Peking University
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt;master at Peking University&lt;/li&gt;
&lt;li&gt;Zhijian Duan, RL, undergraduate at Peking University 
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt; PhD at Peking University&lt;/li&gt;
&lt;/ul&gt;
&lt;h1 id=&#34;2019&#34;&gt;2019&lt;/h1&gt;
&lt;ul&gt;
&lt;li&gt;Mustafa Alghali, AI for social good, master at AMMI
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt;data scientist at Unity&lt;/li&gt;
&lt;li&gt;Ronak Pradeep, Question answering, undergraduate at University of Waterloo
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt;master at University of Waterloo&lt;/li&gt;
&lt;li&gt;&lt;a href=&#34;https://ndai.ai/&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;Ning Dai&lt;/a&gt;, Translation, undergraduate at Fudan University
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt;research intern at ByteDance&lt;/li&gt;
&lt;/ul&gt;
&lt;h1 id=&#34;2018&#34;&gt;2018&lt;/h1&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&#34;https://dayihengliu.github.io/&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;Dayiheng Liu&lt;/a&gt;, Text generation, PhD at Sichuan University
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt;research intern at Microsoft Research Asia&lt;/li&gt;
&lt;li&gt;Shangbang Long, Adversarial attack, undergraduate at Peking University
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt;master at CMU&lt;/li&gt;
&lt;li&gt;&lt;a href=&#34;https://vardaan123.github.io/&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;Vardaan Pahuja&lt;/a&gt;, Visual question answering, master at Mila
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt;PhD at Ohio State University&lt;/li&gt;
&lt;/ul&gt;
&lt;h1 id=&#34;2017&#34;&gt;2017&lt;/h1&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&#34;https://taineleau.me/&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;Danlu Chen&lt;/a&gt;, Hyperparameter optimization, undergraduate at Fudan University
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt;PhD at UCSD&lt;/li&gt;
&lt;li&gt;&lt;a href=&#34;https://www.ritchieng.com/&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;Ritchie Ng&lt;/a&gt;, Hyperparameter optimization, undergraduate at NUS
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt;hedge fund&lt;/li&gt;
&lt;/ul&gt;
&lt;h1 id=&#34;2016&#34;&gt;2016&lt;/h1&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&#34;https://linzichuan.github.io/&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;Zichuan Lin&lt;/a&gt;, RL, undergraduate at Tsinghua University
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt;PhD at Tsinghua University&lt;/li&gt;
&lt;/ul&gt;
&lt;h1 id=&#34;2015&#34;&gt;2015&lt;/h1&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&#34;http://people.csail.mit.edu/hyluo/&#34; target=&#34;_blank&#34; rel=&#34;noopener&#34;&gt;Hongyin Luo&lt;/a&gt;, Hyperparameter optimization, undergraduate at Tsinghua University
  &lt;span class=&#34;inline-block  pr-1&#34;&gt;
    &lt;svg style=&#34;height: 1em; transform: translateY(0.1em);&#34; xmlns=&#34;http://www.w3.org/2000/svg&#34; viewBox=&#34;0 0 24 24&#34;&gt;&lt;path fill=&#34;none&#34; stroke=&#34;currentColor&#34; stroke-linecap=&#34;round&#34; stroke-linejoin=&#34;round&#34; stroke-width=&#34;1.5&#34; d=&#34;M13.5 4.5L21 12m0 0l-7.5 7.5M21 12H3&#34;/&gt;&lt;/svg&gt;
  &lt;/span&gt;PhD at MIT&lt;/li&gt;
&lt;/ul&gt;
</description>
    </item>
    
    <item>
      <title>微信</title>
      <link>/page/chat/</link>
      <pubDate>Sun, 08 Nov 2020 00:00:00 +0000</pubDate>
      <guid>/page/chat/</guid>
      <description>&lt;figure&gt;&lt;img src=&#34;chat.jpg&#34;
    alt=&#34;Please indicate that you find my account through my homepage.&#34;&gt;&lt;figcaption&gt;
      &lt;p&gt;Please indicate that you find my account through my homepage.&lt;/p&gt;
    &lt;/figcaption&gt;
&lt;/figure&gt;

</description>
    </item>
    
    <item>
      <title></title>
      <link>/awards/</link>
      <pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate>
      <guid>/awards/</guid>
      <description></description>
    </item>
    
  </channel>
</rss>
