<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"><channel><description>Assistant professor in Natural Language Processing at the University of Edinburgh and visiting professor at NVIDIA | A Kleene star shines on the hour of our meeting.</description><link>https://bsky.app/profile/edoardo-ponti.bsky.social</link><title>@edoardo-ponti.bsky.social - Edoardo Ponti</title><item><link>https://bsky.app/profile/edoardo-ponti.bsky.social/post/3lqwtsngghs2h</link><description>🚀 By *learning* to compress the KV cache in Transformer LLMs, we can generate more tokens for the same compute budget. &#xA;&#xA;This unlocks *inference-time hyper-scaling*&#xA;&#xA;For the same runtime or memory load, we can boost LLM accuracy by pushing reasoning even further!</description><pubDate>06 Jun 2025 12:33 +0000</pubDate><guid isPermaLink="false">at://did:plc:ctizc4hhwflolzdos4gregjo/app.bsky.feed.post/3lqwtsngghs2h</guid></item><item><link>https://bsky.app/profile/edoardo-ponti.bsky.social/post/3lnnklig7ls2e</link><description>Sparse attention is one of the most promising strategies to unlock long-context processing and long-generation reasoning in LLMs.&#xA;&#xA;We performed the most comprehensive study on training-free sparse attention to date.&#xA;&#xA;Here is what we found:</description><pubDate>25 Apr 2025 15:39 +0000</pubDate><guid isPermaLink="false">at://did:plc:ctizc4hhwflolzdos4gregjo/app.bsky.feed.post/3lnnklig7ls2e</guid></item><item><link>https://bsky.app/profile/edoardo-ponti.bsky.social/post/3lgzycek3cc2s</link><description>I have a scholarship for a PhD in efficient memory and tokenization in LLM architectures at &#xA;@edinburgh-uni.bsky.social!&#xA;&#xA;Eligibility: UK home fee status&#xA;&#xA;Starting date: flexible, from July 2025 onwards.&#xA;&#xA;https://informatics.ed.ac.uk/study-with-us/our-degrees/postgraduate-research-and-cdts/postgraduate-research-funding/phd-efficient-llm-inference&#xA;&#xA;Please contact me if you&#39;re interested!</description><pubDate>31 Jan 2025 12:20 +0000</pubDate><guid isPermaLink="false">at://did:plc:ctizc4hhwflolzdos4gregjo/app.bsky.feed.post/3lgzycek3cc2s</guid></item><item><link>https://bsky.app/profile/edoardo-ponti.bsky.social/post/3lgzx7fdjuk2d</link><description>Code and models for Dynamic Memory Compression are finally available!&#xA;&#xA;Stay tuned for architectures with even more efficient inference.&#xA;&#xA;https://developer.nvidia.com/blog/dynamic-memory-compression/</description><pubDate>31 Jan 2025 12:00 +0000</pubDate><guid isPermaLink="false">at://did:plc:ctizc4hhwflolzdos4gregjo/app.bsky.feed.post/3lgzx7fdjuk2d</guid></item><item><link>https://bsky.app/profile/edoardo-ponti.bsky.social/post/3ldv4zvoovk22</link><description>We&#39;re hiring a lecturer or reader in embodied NLP at the University of Edinburgh!&#xA;&#xA;Deadline: 31 Jan 2025&#xA;Call for applications: https://elxw.fa.em3.oraclecloud.com/hcmUI/CandidateExperience/en/job/11812</description><pubDate>22 Dec 2024 09:46 +0000</pubDate><guid isPermaLink="false">at://did:plc:ctizc4hhwflolzdos4gregjo/app.bsky.feed.post/3ldv4zvoovk22</guid></item><item><link>https://bsky.app/profile/edoardo-ponti.bsky.social/post/3ldra3xd6ps2t</link><description>**Grounded typology**: a new paradigm.&#xA;&#xA;Traditionally, linguists posit functions to compare forms in different languages; however, these are aprioristic and partly arbitrary.&#xA;&#xA;Instead, we resort to perceptual modalities (like vision) as measurable proxies for function.&#xA;&#xA;[contains quote post or other embedded content]</description><pubDate>20 Dec 2024 20:30 +0000</pubDate><guid isPermaLink="false">at://did:plc:ctizc4hhwflolzdos4gregjo/app.bsky.feed.post/3ldra3xd6ps2t</guid></item><item><link>https://bsky.app/profile/edoardo-ponti.bsky.social/post/3ld4t5tlksc2y</link><description>&#xA;Two amazing papers from my students at #NeurIPS today:&#xA;&#xA;⛓️💥 Switch the vocabulary and embeddings of your LLM tokenizer zero-shot on the fly (@bminixhofer.bsky.social)&#xA;https://neurips.cc/virtual/2024/poster/95143&#xA;&#xA;🌊 Align your LLM gradient-free with spectral editing of activations (Yifu Qiu)&#xA;https://neurips.cc/virtual/2024/poster/93529</description><pubDate>12 Dec 2024 17:45 +0000</pubDate><guid isPermaLink="false">at://did:plc:ctizc4hhwflolzdos4gregjo/app.bsky.feed.post/3ld4t5tlksc2y</guid></item><item><link>https://bsky.app/profile/edoardo-ponti.bsky.social/post/3lbyyrsfhkc2d</link><description>We had a blast at this year&#39;s @ellis.eu Dagstuhl seminar on &#34;Modular and Agentive LLMs&#34;. &#xA;&#xA;Thanks everyone for participating!</description><pubDate>28 Nov 2024 11:50 +0000</pubDate><guid isPermaLink="false">at://did:plc:ctizc4hhwflolzdos4gregjo/app.bsky.feed.post/3lbyyrsfhkc2d</guid></item><item><link>https://bsky.app/profile/edoardo-ponti.bsky.social/post/3lbi3m4d5wk2d</link><description>P.S. Make sure to follow @pnawrot.bsky.social!&#xA;&#xA;[contains quote post or other embedded content]</description><pubDate>21 Nov 2024 18:25 +0000</pubDate><guid isPermaLink="false">at://did:plc:ctizc4hhwflolzdos4gregjo/app.bsky.feed.post/3lbi3m4d5wk2d</guid></item><item><link>https://bsky.app/profile/edoardo-ponti.bsky.social/post/3lbhlqzgbcc2a</link><description>Last 5 days to apply for a PhD at #EdinburghNLP!&#xA;&#xA;Deadline: November 25&#xA;&#xA;https://www.ed.ac.uk/studying/postgraduate/degrees/index.php?r=site/view&amp;edition=2025&amp;id=491&#xA;&#xA; If you are passionate about:&#xA;&#xA;- adaptive tokenization and memory in foundation models&#xA;- modular deep learning&#xA;- computational typology&#xA;&#xA;please message me or meet me at #NeurIPS2024!</description><pubDate>21 Nov 2024 13:41 +0000</pubDate><guid isPermaLink="false">at://did:plc:ctizc4hhwflolzdos4gregjo/app.bsky.feed.post/3lbhlqzgbcc2a</guid></item><item><link>https://bsky.app/profile/edoardo-ponti.bsky.social/post/3lbeyh62ga22v</link><description>Another nano gem from my amazing student &#xA;Piotr Nawrot!&#xA;&#xA;A repo &amp; notebook on sparse attention for efficient LLM inference: https://github.com/PiotrNawrot/nano-sparse-attention&#xA;&#xA;This will also feature in my #NeurIPS 2024 tutorial &#34;Dynamic Sparsity in ML&#34; with André Martins: dynamic-sparsity.github.io Stay tuned!</description><pubDate>20 Nov 2024 12:51 +0000</pubDate><guid isPermaLink="false">at://did:plc:ctizc4hhwflolzdos4gregjo/app.bsky.feed.post/3lbeyh62ga22v</guid></item></channel></rss>