<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"><channel><description>Professor at Penn, Amazon Scholar at AWS. Interested in machine learning, uncertainty quantification, game theory, privacy, fairness, and most of the intersections therein</description><link>https://bsky.app/profile/aaroth.bsky.social</link><title>@aaroth.bsky.social - Aaron Roth</title><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mki7bzozqs23</link><description>We updated our paper --- and solved the open problem highlighted in the old version. Now our lower bound construction has only polylog(1/eps) many groups instead of poly(1/eps) many groups. The construction is also simplified.&#xA;&#xA;[contains quote post or other embedded content]</description><pubDate>27 Apr 2026 13:44 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mki7bzozqs23</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mkak6omyt225</link><description>How many samples do you need from an unknown distribution in order to train a model with multicalibration error at most epsilon? &#xA;&#xA;Answer: 1/epsilon^3 samples is both necessary and sufficient.</description><pubDate>24 Apr 2026 12:38 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mkak6omyt225</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mk22hl567k26</link><description>Say hi to @marcelhussing.bsky.social at ICLR&#xA;&#xA;[contains quote post or other embedded content]</description><pubDate>21 Apr 2026 22:40 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mk22hl567k26</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mjmg53ykxs2z</link><description>I&#39;ve recently been getting invitations to talk about how to use AI tools to assist with TCS research. Its something I&#39;ve been doing a lot, but don&#39;t have structured thoughts about how to explain process. But I&#39;m going to try -- first such talk is tomorrow: t.co/wlHPBzXzDm&#xA;https://t.co/wlHPBzXzDm</description><pubDate>16 Apr 2026 12:32 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mjmg53ykxs2z</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mjfniwqerc2p</link><description>AI Agents like Codex are very good at figuring out taxes, including obscure local ones that Intuit doesn&#39;t bother with (looking at you, Philadelphia local taxes). Businesses that provide financial/legal services that involve reasoning through dense but public documentation are in trouble.</description><pubDate>13 Apr 2026 19:55 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mjfniwqerc2p</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mgvnnikwcc2z</link><description>Alpha_0 joke from Dogman</description><pubDate>13 Mar 2026 00:25 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mgvnnikwcc2z</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mguksyjhus2u</link><description>Very cool work. Empirical science has many researcher-degrees-of-freedom which makes it hard to interpret specific studies --- these are only a single trajectory through the data analysis multiverse. Human researchers are opaque. But with agents you can explore the whole space!&#xA;&#xA;[contains quote post or other embedded content]</description><pubDate>12 Mar 2026 14:01 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mguksyjhus2u</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mguhzdqlx22i</link><description>Neural networks are highly non-convex, so approximate error minimizers need not look anything like each other in parameter space. But we show that nevertheless (for many model sizes)  approximate error minimizers must closely agree in function/prediction space despite this!</description><pubDate>12 Mar 2026 13:11 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mguhzdqlx22i</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mgniuktkkk2e</link><description>Michael @mkearnsphilly.bsky.social ) and I wrote a blog post about our experiences using AI for research, and our thoughts on what these developments will mean for research, publication, and education: https://www.amazon.science/blog/how-ai-is-changing-the-nature-of-mathematical-research</description><pubDate>09 Mar 2026 18:38 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mgniuktkkk2e</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mgniuktkkk2e</link><description>Michael @mkearnsphilly.bsky.social ) and I wrote a blog post about our experiences using AI for research, and our thoughts on what these developments will mean for research, publication, and education: https://www.amazon.science/blog/how-ai-is-changing-the-nature-of-mathematical-research</description><pubDate>09 Mar 2026 18:38 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mgniuktkkk2e</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mgauuz2pv22o</link><description>Which is the better model for math? GPT 5.2, or GPT 5.3 codex (either one on high reasoning)?</description><pubDate>04 Mar 2026 18:08 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mgauuz2pv22o</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mbylnwmhi22m</link><description>Excited about a new paper! Multicalibration turns out to be strictly harder than marginal calibration. We prove tight Omega(T^{2/3}) lower bounds for online multicalibration, separating it from online marginal calibration for which better rates were recently discovered.</description><pubDate>09 Jan 2026 13:21 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mbylnwmhi22m</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3mb7mldewh22h</link><description>Yes. We already have a set of ingrained red flags for human written papers that signal a lack of care: not citing the relevant literature, not formatting or typesetting math correctly, etc. These don&#39;t mean the paper is wrong but they strongly correlate with lack of care. But...&#xA;&#xA;[contains quote post or other embedded content]</description><pubDate>30 Dec 2025 15:01 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3mb7mldewh22h</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3majfrvog2s2o</link><description>2025 was an eventful/disruptive year for computer science research, for two reasons: 1) a shock to federal funding, and 2) the arrival of AI models capable enough to assist mathematical research. 1) is unambiguously bad and 2) is probably mostly good. I&#39;ll write about AI first.</description><pubDate>21 Dec 2025 19:01 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3majfrvog2s2o</guid></item><item><link>https://bsky.app/profile/aaroth.bsky.social/post/3m5c7vxj44c2n</link><description>Did your fairness/privacy/CS&amp;Law/etc paper just get rejected from ITCS? Oh FORC!  Submit tomorrow and join us at Harvard this summer.&#xA;&#xA;[contains quote post or other embedded content]</description><pubDate>10 Nov 2025 18:12 +0000</pubDate><guid isPermaLink="false">at://did:plc:3q2kaxhjkceuc7kj4dmtfstl/app.bsky.feed.post/3m5c7vxj44c2n</guid></item></channel></rss>