<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[NLP News]]></title><description><![CDATA[Regular analyses of advances in natural language processing and machine learning.]]></description><link>https://newsletter.ruder.io</link><image><url>https://s3.amazonaws.com/revue/profiles/images/000/027/670/thumb/sebastian_ruder_profile_photo_square.jpg?1616408613</url><title>NLP News</title><link>https://newsletter.ruder.io</link></image><generator>Substack</generator><lastBuildDate>Wed, 08 Apr 2026 14:28:37 GMT</lastBuildDate><atom:link href="https://newsletter.ruder.io/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Sebastian Ruder]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[nlpnewsletter@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[nlpnewsletter@substack.com]]></itunes:email><itunes:name><![CDATA[Sebastian Ruder]]></itunes:name></itunes:owner><itunes:author><![CDATA[Sebastian Ruder]]></itunes:author><googleplay:owner><![CDATA[nlpnewsletter@substack.com]]></googleplay:owner><googleplay:email><![CDATA[nlpnewsletter@substack.com]]></googleplay:email><googleplay:author><![CDATA[Sebastian Ruder]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The Evolving Landscape of LLM Evaluation]]></title><description><![CDATA[Throughout recent years, LLM capabilities have outpaced evaluation benchmarks. This is not a new development. What is new is that the set of standard LLM evals has further narrowed&#8212;and there are questions regarding the reliability of even this small set of benchmarks.]]></description><link>https://newsletter.ruder.io/p/the-evolving-landscape-of-llm-evaluation</link><guid isPermaLink="false">https://newsletter.ruder.io/p/the-evolving-landscape-of-llm-evaluation</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Mon, 13 May 2024 18:54:25 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Edit, May 16: Added mention of <strong>Benchmarking Benchmark Leakage in Large Language Models</strong> (<a href="https://arxiv.org/abs/2404.18824">Xu et al. , 2024)</a>.</p><p>Throughout recent years, LLM capabilities have outpaced evaluation benchmarks. This is not a new development.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> The set of canonical LLM evals has further narrowed to a small set of benchmarks such as <a href="https://arxiv.org/abs/2009.03300">MMLU</a> for general natural language understanding, <a href="https://arxiv.org/abs/2110.14168">GMS8k</a> for mathematical reasoning, and <a href="https://arxiv.org/abs/2107.03374">HumanEval</a> for code, among others. Recently, concerns regarding the reliability of even this small set of benchmarks have emerged.</p><div class="pullquote"><p><em>"Datasets are the telescopes of our field."&#8212;<a href="https://youtu.be/t_A36DDcG_0?t=964&amp;ref=ruder.io">Aravind Joshi</a></em></p></div><p>Without reliable benchmarks, we are effectively flying blind. With many public benchmarks no longer seen as hallmarks of objectivity, ML researchers and practitioners increasingly rely on their intuition to assess a model&#8217;s &#8216;vibe&#8217;, i.e., how interactions with it <em>feel</em> like.</p><p>Let&#8217;s explore how we got here and what the path forward may look like.</p><h3>Looking Back on Benchmarking</h3><p>Benchmarks have been integral to the development of ML and NLP (see <a href="https://www.ruder.io/nlp-benchmarking/#a-brief-history-of-benchmarking">my previous post</a> for a brief history of ML benchmarking). Break-through architectures such as <a href="https://arxiv.org/abs/1512.03385">ResNets</a> or <a href="https://arxiv.org/abs/1706.03762https://arxiv.org/abs/1706.03762">Transformers</a> first captured people&#8217;s attention through their impressive ImageNet and WMT results respectively. Benchmarks such as <a href="https://paperswithcode.com/dataset/mnist">MNIST</a>, <a href="https://paperswithcode.com/dataset/cifar-100">CIFAR-100</a>, and <a href="https://paperswithcode.com/dataset/imagenet">ImageNet</a> have been around for more than a decade and are still used to date.</p><p>In recent years, in light of rapidly improving model capabilities, the time between a benchmark&#8217;s creation to its saturation&#8212;when model performance exceeds human performance&#8212;has dramatically reduced.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!aQpW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ada3dd7-cff8-4c63-b582-7686d31f5e20_2046x984.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!aQpW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ada3dd7-cff8-4c63-b582-7686d31f5e20_2046x984.png 424w, https://substackcdn.com/image/fetch/$s_!aQpW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ada3dd7-cff8-4c63-b582-7686d31f5e20_2046x984.png 848w, https://substackcdn.com/image/fetch/$s_!aQpW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ada3dd7-cff8-4c63-b582-7686d31f5e20_2046x984.png 1272w, https://substackcdn.com/image/fetch/$s_!aQpW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ada3dd7-cff8-4c63-b582-7686d31f5e20_2046x984.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!aQpW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ada3dd7-cff8-4c63-b582-7686d31f5e20_2046x984.png" width="1456" height="700" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7ada3dd7-cff8-4c63-b582-7686d31f5e20_2046x984.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:700,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:202661,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!aQpW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ada3dd7-cff8-4c63-b582-7686d31f5e20_2046x984.png 424w, https://substackcdn.com/image/fetch/$s_!aQpW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ada3dd7-cff8-4c63-b582-7686d31f5e20_2046x984.png 848w, https://substackcdn.com/image/fetch/$s_!aQpW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ada3dd7-cff8-4c63-b582-7686d31f5e20_2046x984.png 1272w, https://substackcdn.com/image/fetch/$s_!aQpW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ada3dd7-cff8-4c63-b582-7686d31f5e20_2046x984.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Benchmark saturation over time for popular benchmarks. Initial performance and human performance are normalised to -1 and 0 respectively (<a href="https://aclanthology.org/2021.naacl-main.324.pdf">Kiela et al., 2021</a>). Performance on MNIST and Switchboard saturated only after 20+ years while performance on GLUE and SQuAD 2.0 saturated already within 1&#8211;2 years.</figcaption></figure></div><p>In addition to recent benchmarks more quickly saturating, two additional problems are contributing to the current benchmark crisis: memorization and overfitting.</p><h3>Memorization</h3><p>Most popular benchmarks are either directly available on the web or may have been uploaded in different forms on GitHub or other platforms. Current models are trained on much of the Internet, with newer models trained on more recent snapshots of CommonCrawl. Unless filtering measures are taken to specifically remove benchmark data from pre-training, models are invariably exposed to test data through their pre-training. While they may not memorize every example, training on the data makes it more likely for the model to produce the correct prediction.</p><p>As a result, LLMs including Aquila2 and Qwen models <a href="https://arxiv.org/abs/2404.18824">repeat verbatim training and even test examples from MATH and GSM8k</a>. GPT models perform much better on coding problems <a href="https://arxiv.org/abs/2310.10628">released before their pre-training data cut-off</a>. LLMs also <a href="https://arxiv.org/abs/2312.16337">perform much better on datasets released before the pre-training data cut-off</a>&#8212;and barely improve over a majority baseline (!) on uncontaminated classification tasks in zero and few-shot settings<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>.</p><p>What can we do to mitigate memorization? <a href="https://arxiv.org/abs/2404.00699">Ravaut et al. (2024)</a> highlight some best practices to reduce contamination in their survey:</p><ol><li><p><strong>Encrypting evaluation datasets</strong>. The authors of the <a href="https://arxiv.org/abs/2309.16575">MTOB</a> dataset (discussed in a <a href="https://newsletter.ruder.io/p/true-zero-shot-mt">previous newsletter</a>) did this. The downside: if the dataset is shared unencrypted at any point, it is hard to contain it again.</p></li><li><p><strong>Scanning newly released evaluation datasets</strong>. We can make sure to only evaluate on uncontaminated data.</p></li><li><p><strong>Preventing data leakage to closed-source APIs</strong>. Evaluating on closed-source APIs inadvertently leaks data to them whether or not the data is made available online or not.</p></li></ol><h3>Overfitting</h3><p>With increased attention afforded to LLMs and billions of dollars in funding at play, the pressure to do well on public benchmarks has increased. A couple percentage points on an established benchmark such as <a href="https://arxiv.org/pdf/2009.03300">MMLU</a> can make or break an investor presentation or convince potential customers to try a model.</p><p>As a result, there is the risk of overfitting to public benchmarks if models are optimized to do well on them. One way of overfitting is through the creation of synthetic data, which may inadvertently reflect use cases in the test data rather than a broader set of model applications. For example, <a href="https://arxiv.org/abs/2405.00332">a recent study</a> found that several model families such as Phi and Mistral models show evidence of systematic overfitting on the GSM8k grade school math dataset.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Hmpz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Hmpz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png 424w, https://substackcdn.com/image/fetch/$s_!Hmpz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png 848w, https://substackcdn.com/image/fetch/$s_!Hmpz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png 1272w, https://substackcdn.com/image/fetch/$s_!Hmpz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Hmpz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png" width="1456" height="785" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:785,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:731523,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Hmpz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png 424w, https://substackcdn.com/image/fetch/$s_!Hmpz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png 848w, https://substackcdn.com/image/fetch/$s_!Hmpz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png 1272w, https://substackcdn.com/image/fetch/$s_!Hmpz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe92dc248-5d16-43b1-81db-02616bce3e58_2384x1286.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Models arranged by their drop in performance between GSM8k and the newly created GSM1k (lower is worse). Mistral and Phi models show a drop of 10% on GSM1k compared to GSM8k (<a href="https://arxiv.org/abs/2405.00332">Zhang et al., 2024</a>).</figcaption></figure></div><p>For benchmarks such as <a href="https://arxiv.org/abs/2306.05685">MT-Bench</a> that employ a specific model (often GPT4) as evaluator, there is the additional risk of overfitting to biases of the evaluator. If we optimize our model to do well on MT-Bench, we may end up training on GPT4-created data. This can lead to higher scores on GPT4-rated evals but worse performance when tested by humans as models <a href="https://arxiv.org/abs/2305.15717">mimic GPT4&#8217;s style but not other aspects such as its factuality</a>.</p><p>Training on such model-created synthetic data may thus lead to gains and improve evals in the short term; in the long term, it can lead to unanticipated biases and blind spots in the user experience. </p><p>Taking into account the possibility of both memorization and overfitting, the results on popular static evaluation benchmarks should be taken with a grain of salt.</p><h2>To Vibe or Not to Vibe?</h2><p>Instead of blindly trusting public benchmarks or result tables in press releases, it is thus more important than ever to run your own tests. LLMs are versatile and different people have different preferences and use cases. <a href="https://aclanthology.org/2020.emnlp-main.393/">Utility is in the eye of the user</a>. However, very few have the means to exhaustively evaluate and compare many different LLMs.</p><p>The currently preferred evaluation benchmark for such &#8216;vibe-based&#8217; evals is <a href="https://chat.lmsys.org/">Chatbot Arena</a>, which crowd-sources user ratings in blind A/B test conversations with various LLMs. The platform enables large-scale user testing and aggregates win rates across 10,000s of conversations in a central leaderboard.</p><p>Chatbot Arena is not perfect. Humans can get fooled and may prefer a response due to a variety of factors including its formatting, style, type of humor, etc, which can be exploited and optimized for. Chatbot Arena covers a narrow range of domains and conversations wildly vary in quality. Nevertheless, it provides an uncontaminated evaluation of chat user interactions (I&#8217;ve <a href="https://newsletter.ruder.io/p/command-r#%C2%A7chatbot-arena">written previously</a> about it); you can also check out <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Nathan Lambert&quot;,&quot;id&quot;:10472909,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fdda47b96-836a-4b95-99a0-f0ec744d4245_2316x2316.jpeg&quot;,&quot;uuid&quot;:&quot;061c2272-2614-43bc-b7ed-75ff918ac374&quot;}" data-component-name="MentionToDOM"></span>&#8217;s thoughts on Chatbot Arena in <a href="https://www.interconnects.ai/p/chatbotarena-the-future-of-llm-evaluation">his recent post</a>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">NLP News is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>The Future of Evaluation</h2><p>The time when benchmarks lasted multiple decades has passed. Going forward, we will rely less on public benchmark results. Instead, the <strong>ability to evaluate a model directly for a specific downstream use case will be much more important</strong>.</p><p>This requires <strong>a) knowledge on how to efficiently and robustly evaluate LLMs</strong>; <strong>b) the infrastructure</strong> to enable this; and <strong>c) domain expertise</strong> to know how the problem can be modeled using an LLM. Institutions will be less likely to share such evals given that a release means they will become contaminated, reducing their usefulness.</p><p><strong>Benchmark creators should mitigate the risk of contamination</strong> in the design of their test data as much as possible. <strong>Internet data should be used sparingly</strong>&#8212;and not as the source of the solution. <strong>New benchmark data should ideally be created by humans</strong> from scratch.</p><p>As we already evaluate LLMs how we assess humans&#8212;on standardized exams and general aptitude and proficiency tests&#8212;<strong>we should also follow a similar design process for future LLM evals: With regularly updated tests, assuming access to all prior exam data and accounting for a willingness to exploit loopholes</strong>.</p><p>Overall, we need to rethinking the way we evaluate LLMs and shift to efficient evaluation processes that can keep track with the pace of model advances. </p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The <a href="https://aiindex.stanford.edu/wp-content/uploads/2021/11/2021-AI-Index-Report_Master.pdf">AI Index Report 2021</a> mentions this and I&#8217;ve previously written about <a href="https://www.ruder.io/nlp-benchmarking/">challenges in NLP benchmarking</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>The authors examined less powerful LLMs up to GPT3.5-Turbo (March 2023 release). It is likely that more powerful LLMs would exhibit stronger zero-shot and few-shot capabilities.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Command R+]]></title><description><![CDATA[The Top Open-Weights LLM + RAG and Multilingual Support]]></description><link>https://newsletter.ruder.io/p/command-r</link><guid isPermaLink="false">https://newsletter.ruder.io/p/command-r</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Mon, 15 Apr 2024 17:23:52 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5a71de9d-6d4e-4c69-b464-1ae8ebbc20d9_2000x1125.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This post is an update on what I&#8217;ve been up to since I joined Cohere. I&#8217;ve had fun contributing to the launch of <a href="https://txt.cohere.com/command-r/">Command R</a> and <a href="https://txt.cohere.com/command-r-plus-microsoft-azure/">R+</a>, the latest Cohere models. I&#8217;ll discuss more details once the tech report is out; in the meantime I&#8217;ll share what I&#8217;m most excited about.</p><p>Command R+ is ranked as the top open-weights model on <a href="https://chat.lmsys.org/?leaderboard">Chatbot Arena</a>, even outperforming some versions of GPT-4. Why is this exciting?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QldV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46722be1-5670-43e9-9d5e-f24fced20621_1756x694.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QldV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46722be1-5670-43e9-9d5e-f24fced20621_1756x694.jpeg 424w, https://substackcdn.com/image/fetch/$s_!QldV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46722be1-5670-43e9-9d5e-f24fced20621_1756x694.jpeg 848w, https://substackcdn.com/image/fetch/$s_!QldV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46722be1-5670-43e9-9d5e-f24fced20621_1756x694.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!QldV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46722be1-5670-43e9-9d5e-f24fced20621_1756x694.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QldV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46722be1-5670-43e9-9d5e-f24fced20621_1756x694.jpeg" width="1456" height="575" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/46722be1-5670-43e9-9d5e-f24fced20621_1756x694.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:575,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:209892,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QldV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46722be1-5670-43e9-9d5e-f24fced20621_1756x694.jpeg 424w, https://substackcdn.com/image/fetch/$s_!QldV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46722be1-5670-43e9-9d5e-f24fced20621_1756x694.jpeg 848w, https://substackcdn.com/image/fetch/$s_!QldV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46722be1-5670-43e9-9d5e-f24fced20621_1756x694.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!QldV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46722be1-5670-43e9-9d5e-f24fced20621_1756x694.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Chatbot Arena leaderboard as of April 9, 2024 (source: <a href="https://x.com/lmsysorg/status/1777630133798772766">lmsys.org</a>).</figcaption></figure></div><h2><strong>Chatbot Arena</strong></h2><p>Let&#8217;s talk about why we should care about Chatbot Arena rankings in the first place. I&#8217;ve written in the past about <a href="https://www.ruder.io/nlp-benchmarking/">challenges in NLP benchmarking</a>. Pre-LLM benchmarks such as <a href="https://super.gluebenchmark.com/">SuperGLUE</a> mostly consist of classification tasks and no longer provide sufficient signal to differentiate the latest generation of LLMs. More recent benchmarks such as <a href="https://arxiv.org/abs/2306.05685">MT-Bench</a> consist of small samples of open-ended questions and rely on LLMs as evaluators, which have their <a href="https://arxiv.org/abs/2305.17926">own sets</a> <a href="https://arxiv.org/abs/2307.03025">of biases</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p><a href="https://arxiv.org/abs/2009.03300">MMLU</a>, one of the most widely used benchmarks consisting of 14k multiple-choice questions sourced from public sources covering 57 domains has been featured prominently in <a href="https://openai.com/research/gpt-4">GPT-4</a>, <a href="https://www.anthropic.com/news/claude-3-family">Claude 3</a>, and <a href="https://mistral.ai/news/mistral-large/">Mistral Large</a> posts. The data is <a href="https://www.youtube.com/watch?v=hVade_8H8mE">not without errors</a>, however, and given its release in 2020, training data of recent models is likely at <a href="https://arxiv.org/abs/2310.10628">least partially contaminated</a>.</p><p><a href="https://chat.lmsys.org/">Chatbot Arena</a> is a platform where users rate conversations in a blind A/B test. They can continue the conversation until they choose a winner. Of course, short user interactions often do not reveal more advanced model capabilities and annotators can be <a href="https://arxiv.org/abs/2305.15717">fooled by authoritative but non-factual answers</a>. Nevertheless, this is the closest to an assessment on realistic user interactions that we currently have. As models are always evaluated based on new user conversations, there is no risk of data contamination.</p><p>Command R+ outperforms versions of GPT-4 on Chatbot Arena while being much cheaper to use. It does also well on use cases that are under-represented in Chatbot Arena such as RAG, tool use, and multilinguality.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WAOP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e12e01-1fd0-451e-ae6c-5e4a6a0ceeab_2595x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WAOP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e12e01-1fd0-451e-ae6c-5e4a6a0ceeab_2595x900.png 424w, https://substackcdn.com/image/fetch/$s_!WAOP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e12e01-1fd0-451e-ae6c-5e4a6a0ceeab_2595x900.png 848w, https://substackcdn.com/image/fetch/$s_!WAOP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e12e01-1fd0-451e-ae6c-5e4a6a0ceeab_2595x900.png 1272w, https://substackcdn.com/image/fetch/$s_!WAOP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e12e01-1fd0-451e-ae6c-5e4a6a0ceeab_2595x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WAOP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e12e01-1fd0-451e-ae6c-5e4a6a0ceeab_2595x900.png" width="1456" height="505" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/80e12e01-1fd0-451e-ae6c-5e4a6a0ceeab_2595x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:505,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:73498,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WAOP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e12e01-1fd0-451e-ae6c-5e4a6a0ceeab_2595x900.png 424w, https://substackcdn.com/image/fetch/$s_!WAOP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e12e01-1fd0-451e-ae6c-5e4a6a0ceeab_2595x900.png 848w, https://substackcdn.com/image/fetch/$s_!WAOP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e12e01-1fd0-451e-ae6c-5e4a6a0ceeab_2595x900.png 1272w, https://substackcdn.com/image/fetch/$s_!WAOP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80e12e01-1fd0-451e-ae6c-5e4a6a0ceeab_2595x900.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">(left) Performance comparison of Command R+, Mistral-Large, and GPT4-turbo on three key capabilities: Multilingual, RAG, and Tool Use. (right) Comparison input and output token costs per million for models available on Azure. <a href="https://txt.cohere.com/command-r-plus-microsoft-azure/">Source</a></figcaption></figure></div><h2><strong>A GPT-4 Level Model on Your Computer</strong></h2><p>Command R+ consists of 104B parameters with <a href="https://huggingface.co/CohereForAI/c4ai-command-r-plus">publicly available weights</a>. This is the first time that a model that is close to GPT-4 performance is available for research use. With the right setup, Command R+ can generate text at a <a href="https://x.com/huggingface/status/1778847929295679713">rate of 111 tokens/s</a> (!) when deployed locally.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> To understand how to effectively prompt the model, check out the <a href="https://docs.cohere.com/docs/prompting-command-r">prompting guide</a>.</p><p>I&#8217;m excited about what this means for the open-source community and research, with the gap between closed-source and open-weight models closing and SOTA-level conversational models being more easily accessible.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!A6sk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc31cece-3494-44fa-b08f-21638bb7d28a_1158x764.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!A6sk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc31cece-3494-44fa-b08f-21638bb7d28a_1158x764.jpeg 424w, https://substackcdn.com/image/fetch/$s_!A6sk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc31cece-3494-44fa-b08f-21638bb7d28a_1158x764.jpeg 848w, https://substackcdn.com/image/fetch/$s_!A6sk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc31cece-3494-44fa-b08f-21638bb7d28a_1158x764.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!A6sk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc31cece-3494-44fa-b08f-21638bb7d28a_1158x764.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!A6sk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc31cece-3494-44fa-b08f-21638bb7d28a_1158x764.jpeg" width="1158" height="764" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dc31cece-3494-44fa-b08f-21638bb7d28a_1158x764.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:764,&quot;width&quot;:1158,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:88239,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!A6sk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc31cece-3494-44fa-b08f-21638bb7d28a_1158x764.jpeg 424w, https://substackcdn.com/image/fetch/$s_!A6sk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc31cece-3494-44fa-b08f-21638bb7d28a_1158x764.jpeg 848w, https://substackcdn.com/image/fetch/$s_!A6sk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc31cece-3494-44fa-b08f-21638bb7d28a_1158x764.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!A6sk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc31cece-3494-44fa-b08f-21638bb7d28a_1158x764.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The gap between closed-source and open-weights models on Chatbot Arena is closing (source: <a href="https://x.com/maximelabonne/status/1779801605702836454">Maxime Labonne</a>).</figcaption></figure></div><p>Other recently released models such as <a href="https://huggingface.co/databricks/dbrx-instruct">DBRX</a> (132B parameters), <a href="https://huggingface.co/mistral-community/Mixtral-8x22B-v0.1-4bit">Mixtral 8x22B</a> (176B parameters), and <a href="https://x.ai/blog/grok-os">Grok-1</a> (314B parameters) are based on a Mixture-of-Experts (MoE), trading off inference speed for memory costs. While these models only activate a subset of parameters for each token, they still require storing all parameters in-memory, which makes them harder to use locally. So far, they are not available or rank much lower than Command R+ on <a href="https://chat.lmsys.org/">Chatbot Arena</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>Command R+ comes with a non-commercial license. If you want to self-host or fine-tune it for commercial purposes, <a href="https://x.com/aidangomez/status/1775878627231515049">we&#8217;ll work with you</a> to find something that works for you.</p><h2>RAG and Tool Use</h2><p>While Command R+ can be used as a chatbot, it has been designed for enterprise use. Faithful and verifiable responses are especially important in an enterprise setting. Reducing hallucinations and providing trustworthy responses are important research challenges. There are different ways to mitigate hallucinations, ranging from debiasing and model editing to specialized decoding strategies (see <a href="https://arxiv.org/abs/2311.05232">Huang et al. (2023)</a> for an overview).</p><p>Retrieval-augmented generation (RAG; <a href="https://arxiv.org/abs/2005.11401">Lewis et al., 2020</a>), which conditions on the LLM&#8217;s generation on retrieved documents is the most practical paradigm IMO. Command R+ uses RAG with in-line citations to provide grounded responses.</p><p>However, evaluation of the quality and trustworthiness of such responses is challenging and motivated the development of new evaluation frameworks such as Attributable to Identified Sources (AIS; <a href="https://direct.mit.edu/coli/article/49/4/777/116438/Measuring-Attribution-in-Natural-Language">Rashkin et al., 2023</a>).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> On our internal human evaluation measuring citation fidelity, Command R+ outperforms GPT4-turbo. On public multi-hop QA benchmarks, it outperforms models at the same price point such as Claude 3 Sonnet and Mistral-large.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ykm3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf8c88bf-3b02-46da-b97d-a665ba80d9ee_2365x950.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ykm3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf8c88bf-3b02-46da-b97d-a665ba80d9ee_2365x950.png 424w, https://substackcdn.com/image/fetch/$s_!ykm3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf8c88bf-3b02-46da-b97d-a665ba80d9ee_2365x950.png 848w, https://substackcdn.com/image/fetch/$s_!ykm3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf8c88bf-3b02-46da-b97d-a665ba80d9ee_2365x950.png 1272w, https://substackcdn.com/image/fetch/$s_!ykm3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf8c88bf-3b02-46da-b97d-a665ba80d9ee_2365x950.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ykm3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf8c88bf-3b02-46da-b97d-a665ba80d9ee_2365x950.png" width="1456" height="585" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/af8c88bf-3b02-46da-b97d-a665ba80d9ee_2365x950.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:585,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:76732,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ykm3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf8c88bf-3b02-46da-b97d-a665ba80d9ee_2365x950.png 424w, https://substackcdn.com/image/fetch/$s_!ykm3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf8c88bf-3b02-46da-b97d-a665ba80d9ee_2365x950.png 848w, https://substackcdn.com/image/fetch/$s_!ykm3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf8c88bf-3b02-46da-b97d-a665ba80d9ee_2365x950.png 1272w, https://substackcdn.com/image/fetch/$s_!ykm3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf8c88bf-3b02-46da-b97d-a665ba80d9ee_2365x950.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">(left) Human head-to-head preference results using a holistic grading scheme combining text fluency, citation quality, and overall utility. (right) Accuracy of multi-hop REACT agents powered by various models with access to the same search tools retrieving from Wikipedia (HotpotQA) and the Internet (Bamboogle and StrategyQA). <a href="https://txt.cohere.com/command-r-plus-microsoft-azure/">Source</a></figcaption></figure></div><p>You can easily use RAG via the <a href="https://docs.cohere.com/docs/retrieval-augmented-generation-rag">API</a> on the Internet or your own documents. A complete RAG workflow additionally involves document search and reranking, which can be seen in <a href="https://github.com/cohere-ai/notebooks/blob/main/notebooks/Vanilla_RAG.ipynb">this Colab</a> with an example RAG setup on Wikipedia.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kpb4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c39a636-dd39-4429-86f7-a43307ef0d11_2524x750.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kpb4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c39a636-dd39-4429-86f7-a43307ef0d11_2524x750.png 424w, https://substackcdn.com/image/fetch/$s_!kpb4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c39a636-dd39-4429-86f7-a43307ef0d11_2524x750.png 848w, https://substackcdn.com/image/fetch/$s_!kpb4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c39a636-dd39-4429-86f7-a43307ef0d11_2524x750.png 1272w, https://substackcdn.com/image/fetch/$s_!kpb4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c39a636-dd39-4429-86f7-a43307ef0d11_2524x750.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kpb4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c39a636-dd39-4429-86f7-a43307ef0d11_2524x750.png" width="1456" height="433" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8c39a636-dd39-4429-86f7-a43307ef0d11_2524x750.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:433,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:197798,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kpb4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c39a636-dd39-4429-86f7-a43307ef0d11_2524x750.png 424w, https://substackcdn.com/image/fetch/$s_!kpb4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c39a636-dd39-4429-86f7-a43307ef0d11_2524x750.png 848w, https://substackcdn.com/image/fetch/$s_!kpb4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c39a636-dd39-4429-86f7-a43307ef0d11_2524x750.png 1272w, https://substackcdn.com/image/fetch/$s_!kpb4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c39a636-dd39-4429-86f7-a43307ef0d11_2524x750.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Example of basic RAG usage with Command models using the Cohere API (<a href="https://docs.cohere.com/docs/retrieval-augmented-generation-rag">source</a>).</figcaption></figure></div><p>In enterprise settings, seamless integrations with existing APIs and services is crucial. I&#8217;ve <a href="https://newsletter.ruder.io/p/tool-augmented-llms">written before</a> about the promise of tool-augmented models. Tools can help decompose complex problems and make LLMs outputs more interpretable by enabling users to look at the trace of API calls. Command R+ has been trained for zero-shot multi-step tool use. On public tool use benchmarks, it outperforms GPT4-turbo.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lJP4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5650f97-8118-4abc-903d-6046ee1981ac_1400x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lJP4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5650f97-8118-4abc-903d-6046ee1981ac_1400x900.png 424w, https://substackcdn.com/image/fetch/$s_!lJP4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5650f97-8118-4abc-903d-6046ee1981ac_1400x900.png 848w, https://substackcdn.com/image/fetch/$s_!lJP4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5650f97-8118-4abc-903d-6046ee1981ac_1400x900.png 1272w, https://substackcdn.com/image/fetch/$s_!lJP4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5650f97-8118-4abc-903d-6046ee1981ac_1400x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lJP4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5650f97-8118-4abc-903d-6046ee1981ac_1400x900.png" width="1400" height="900" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c5650f97-8118-4abc-903d-6046ee1981ac_1400x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:900,&quot;width&quot;:1400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:48283,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lJP4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5650f97-8118-4abc-903d-6046ee1981ac_1400x900.png 424w, https://substackcdn.com/image/fetch/$s_!lJP4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5650f97-8118-4abc-903d-6046ee1981ac_1400x900.png 848w, https://substackcdn.com/image/fetch/$s_!lJP4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5650f97-8118-4abc-903d-6046ee1981ac_1400x900.png 1272w, https://substackcdn.com/image/fetch/$s_!lJP4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5650f97-8118-4abc-903d-6046ee1981ac_1400x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Conversational tool-use and single-turn function-calling evaluations using Microsoft&#8217;s ToolTalk (Hard) benchmark (<a href="https://arxiv.org/abs/2311.10775?ref=txt.cohere.com">Farn &amp; Shin 2023</a>) and Berkeley's Function Calling Leaderboard (BFCL) (<a href="https://gorilla.cs.berkeley.edu/blogs/8_berkeley_function_calling_leaderboard.html?ref=txt.cohere.com">Yan et al. 2024</a>). <a href="https://txt.cohere.com/command-r-plus-microsoft-azure/">Source</a></figcaption></figure></div><p>The recommended way to leverage multi-step tool use with Command R+ is via <a href="https://python.langchain.com/docs/integrations/providers/cohere?ref=txt.cohere.com">LangChain</a>. To teach the model to use a new tool, you only need to provide the name, definition (a Python function), and the arguments schema. The model can then be used as a ReAct agent in LangChain with a range of tools (see <a href="https://github.com/cohere-ai/notebooks/blob/main/notebooks/Vanilla_Multi_Step_Tool_Use.ipynb">this Colab</a> for an example workflow).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vZYi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a9dee15-db43-412f-a865-a5184b2db28e_2504x1338.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vZYi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a9dee15-db43-412f-a865-a5184b2db28e_2504x1338.png 424w, https://substackcdn.com/image/fetch/$s_!vZYi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a9dee15-db43-412f-a865-a5184b2db28e_2504x1338.png 848w, https://substackcdn.com/image/fetch/$s_!vZYi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a9dee15-db43-412f-a865-a5184b2db28e_2504x1338.png 1272w, https://substackcdn.com/image/fetch/$s_!vZYi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a9dee15-db43-412f-a865-a5184b2db28e_2504x1338.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vZYi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a9dee15-db43-412f-a865-a5184b2db28e_2504x1338.png" width="1456" height="778" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8a9dee15-db43-412f-a865-a5184b2db28e_2504x1338.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:778,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:298507,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vZYi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a9dee15-db43-412f-a865-a5184b2db28e_2504x1338.png 424w, https://substackcdn.com/image/fetch/$s_!vZYi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a9dee15-db43-412f-a865-a5184b2db28e_2504x1338.png 848w, https://substackcdn.com/image/fetch/$s_!vZYi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a9dee15-db43-412f-a865-a5184b2db28e_2504x1338.png 1272w, https://substackcdn.com/image/fetch/$s_!vZYi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a9dee15-db43-412f-a865-a5184b2db28e_2504x1338.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Example of using Command R+ in LangChain with different tools including Internet search, vector store search, and Python execution.</figcaption></figure></div><p>I hope that strong support for RAG and tool use in an open-weights model will lead to progress in important research directions, some of which I have outlined <a href="https://newsletter.ruder.io/i/136248186/the-future-of-tool-augmented-llms">here</a>. If you want the most efficient solution for RAG, <a href="https://txt.cohere.com/command-r/">Command R</a> demonstrates highly competitive RAG and tool-use performance at cheaper cost (35B-parameter weights are <a href="https://huggingface.co/CohereForAI/c4ai-command-r-v01">publicly available</a>).</p><h2>Multilingual</h2><p>Command R+ works well in languages beyond English. It was <a href="https://huggingface.co/CohereForAI/c4ai-command-r-plus#model-details">pre-trained on 23 languages</a>, with our main focus on 10 key language of global business: English, French, Spanish, Italian, German, Brazilian Portuguese, Japanese, Korean, Simplified Chinese, and Arabic. In our evaluations on translation tasks, it is competitive with GPT4-turbo.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ziht!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904978a1-54c2-40d8-9bd0-2dc4a1481618_1400x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ziht!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904978a1-54c2-40d8-9bd0-2dc4a1481618_1400x900.png 424w, https://substackcdn.com/image/fetch/$s_!Ziht!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904978a1-54c2-40d8-9bd0-2dc4a1481618_1400x900.png 848w, https://substackcdn.com/image/fetch/$s_!Ziht!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904978a1-54c2-40d8-9bd0-2dc4a1481618_1400x900.png 1272w, https://substackcdn.com/image/fetch/$s_!Ziht!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904978a1-54c2-40d8-9bd0-2dc4a1481618_1400x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ziht!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904978a1-54c2-40d8-9bd0-2dc4a1481618_1400x900.png" width="1400" height="900" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/904978a1-54c2-40d8-9bd0-2dc4a1481618_1400x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:900,&quot;width&quot;:1400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:46475,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ziht!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904978a1-54c2-40d8-9bd0-2dc4a1481618_1400x900.png 424w, https://substackcdn.com/image/fetch/$s_!Ziht!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904978a1-54c2-40d8-9bd0-2dc4a1481618_1400x900.png 848w, https://substackcdn.com/image/fetch/$s_!Ziht!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904978a1-54c2-40d8-9bd0-2dc4a1481618_1400x900.png 1272w, https://substackcdn.com/image/fetch/$s_!Ziht!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904978a1-54c2-40d8-9bd0-2dc4a1481618_1400x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Comparison of models on FLoRES (in French, Spanish, Italian, German, Portuguese, Japanese, Korean, Arabic, and Chinese) and WMT23 (in German, Japanese, and Chinese) translation tasks.</figcaption></figure></div><p>Command R+ has been designed with multilinguality in mind. Its tokenizer is much less English-centric than others and compresses text in non-English languages much better than both the Mistral and OpenAI tokenizers.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a> As LLM providers charge based on the number of input/output tokens, tokenizer choice directly impacts API costs for users. At the same cost-per-token, if one LLM generates 2x as many tokens as another, the API costs will also be twice as large.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ezBL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b3379b6-11fd-40ff-8cb5-74e347e6ed48_1740x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ezBL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b3379b6-11fd-40ff-8cb5-74e347e6ed48_1740x900.png 424w, https://substackcdn.com/image/fetch/$s_!ezBL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b3379b6-11fd-40ff-8cb5-74e347e6ed48_1740x900.png 848w, https://substackcdn.com/image/fetch/$s_!ezBL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b3379b6-11fd-40ff-8cb5-74e347e6ed48_1740x900.png 1272w, https://substackcdn.com/image/fetch/$s_!ezBL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b3379b6-11fd-40ff-8cb5-74e347e6ed48_1740x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ezBL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b3379b6-11fd-40ff-8cb5-74e347e6ed48_1740x900.png" width="1456" height="753" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4b3379b6-11fd-40ff-8cb5-74e347e6ed48_1740x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:753,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:45225,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ezBL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b3379b6-11fd-40ff-8cb5-74e347e6ed48_1740x900.png 424w, https://substackcdn.com/image/fetch/$s_!ezBL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b3379b6-11fd-40ff-8cb5-74e347e6ed48_1740x900.png 848w, https://substackcdn.com/image/fetch/$s_!ezBL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b3379b6-11fd-40ff-8cb5-74e347e6ed48_1740x900.png 1272w, https://substackcdn.com/image/fetch/$s_!ezBL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4b3379b6-11fd-40ff-8cb5-74e347e6ed48_1740x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Comparison of the number of tokens produced by the Cohere, Mistral (Mixtral), and OpenAI tokenizers for different languages (as a multiple of the number of tokens produced by the Cohere tokenizer). The Cohere tokenizer produces much fewer tokens to represent the same text, with particularly large reductions on non-Latin script languages. For instance, in Japanese, the OpenAI tokenizer outputs 1.67x as many tokens as the Cohere tokenizer.</figcaption></figure></div><p><a href="https://aclanthology.org/2023.emnlp-main.614/">Ahia et al. (2023)</a> highlighted that such over-segmentation leads to &#8220;double unfairness&#8221;: higher API prices and lower utility (reduced performance) for many languages. In comparison, Command R+ is much more equitable. I hope that companies will take into account the impact of tokenization and other design choices on API costs in future LLMs.</p><p>Given the focus on the Latin script in existing models, I particularly want to highlight Command R+&#8217;s performance in some prominent non-Latin script languages: Japanese, Korean, and Chinese. We evaluated on translation tasks as well as language-specific benchmarks such as <a href="https://github.com/Stability-AI/FastChat">Japanese MT-Bench</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>. Command R+ outperforms Claude 3 Sonnet and Mistral Large<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> and is competitive with GPT 4-Turbo.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2z8z!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc0d6b57-62da-4742-b6b9-3981773a0f21_1600x679.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2z8z!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc0d6b57-62da-4742-b6b9-3981773a0f21_1600x679.jpeg 424w, https://substackcdn.com/image/fetch/$s_!2z8z!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc0d6b57-62da-4742-b6b9-3981773a0f21_1600x679.jpeg 848w, https://substackcdn.com/image/fetch/$s_!2z8z!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc0d6b57-62da-4742-b6b9-3981773a0f21_1600x679.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!2z8z!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc0d6b57-62da-4742-b6b9-3981773a0f21_1600x679.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2z8z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc0d6b57-62da-4742-b6b9-3981773a0f21_1600x679.jpeg" width="1456" height="618" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dc0d6b57-62da-4742-b6b9-3981773a0f21_1600x679.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:618,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:62899,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2z8z!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc0d6b57-62da-4742-b6b9-3981773a0f21_1600x679.jpeg 424w, https://substackcdn.com/image/fetch/$s_!2z8z!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc0d6b57-62da-4742-b6b9-3981773a0f21_1600x679.jpeg 848w, https://substackcdn.com/image/fetch/$s_!2z8z!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc0d6b57-62da-4742-b6b9-3981773a0f21_1600x679.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!2z8z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc0d6b57-62da-4742-b6b9-3981773a0f21_1600x679.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Japanese evaluation on FLoRES, WMT23, and Japanese MT-Bench (<a href="https://x.com/seb_ruder/status/1778386288313315833">source</a>).</figcaption></figure></div><p>We see similar trends for evaluations in <a href="https://x.com/seb_ruder/status/1778386446493135058">Korean</a> and <a href="https://x.com/seb_ruder/status/1778386610201014339">Chinese</a>. On <a href="https://x.com/lmsysorg/status/1778555685598568946">Chinese Chatbot Arena</a>, Command R+ is only behind GPT4 and Claude 3 Opus, models that are 2&#8211;3x more expensive. It&#8217;s been exciting to read the feedback from speakers of different language communities using Command R+.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">NLP News is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Conclusion</h2><p>Overall, I&#8217;m really excited about Command R+&#8217;s capabilities and the future of LLMs. We will be pushing its multilingual capabilities to make it useful in many languages used in business. I&#8217;ve been using Command R+ with RAG via <a href="https://dashboard.cohere.com/playground/chat">Cohere&#8217;s playground</a> for my exploratory Internet searches and creative tasks in English and German and have been impressed by its quality. Feel free to try it in your language&#8212;and share your feedback in the comments or via email. I&#8217;d love to hear what works or doesn&#8217;t work for you.</p><p>I&#8217;m also excited to hear from you if you&#8217;d like to explore using Command R+ for your (multilingual) business applications.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>LLM evaluators prefer longer responses or are affected by the order in which responses are presented.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Evaluation results on other public benchmarks can be found <a href="https://huggingface.co/CohereForAI/c4ai-command-r-plus#evaluations">here</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>This is on 4x A100 GPUs using a highly optimized open-source backend. Quantized versions of Command R+ can be run on 2x 3090 GPUs at <a href="https://www.reddit.com/r/LocalLLaMA/comments/1c3stb8/comment/kzlj0kk/?utm_source=share&amp;utm_medium=web3x&amp;utm_name=web3xcss&amp;utm_term=1&amp;utm_content=share_button">around 10 tokens/s</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>DBRX ranks at #26 on Chatbot Arena as of the publication date of this post.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Things used to be easier when the default setting was purely extractive QA and accuracy and exact match (EM) were the go-to metrics (see, for instance, <a href="https://research.google/pubs/natural-questions-a-benchmark-for-question-answering-research/">Natural Questions</a> or <a href="https://arxiv.org/abs/2003.05002">TyDi QA</a>). With generative models, automatically identifying whether a (possibly verbose) response answers a question is much more challenging.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Evaluation for HotpotQA and Bamboogle is done via a committee-of-LLMs-as-judges to reduce evaluator bias.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>The Anthropic tokenizer is not public so we could not compare to them.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>We use GPT4 as an evaluator so this evaluation is biased towards GPT4-turbo.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>Mistral Large doesn&#8217;t officially support these languages so its results are expectedly lower.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[True Zero-shot MT]]></title><description><![CDATA[Teaching Machines a New Language Like Humans]]></description><link>https://newsletter.ruder.io/p/true-zero-shot-mt</link><guid isPermaLink="false">https://newsletter.ruder.io/p/true-zero-shot-mt</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Tue, 27 Feb 2024 09:46:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Little over a week ago, <a href="https://x.com/JeffDean/status/1758149033473020081?s=20">Gemini 1.5 reported</a> close to human-level performance on <a href="https://arxiv.org/abs/2309.16575">MTOB</a>, a recent challenging translation dataset. In this newsletter, we&#8217;ll dig into this result, explore true zero-shot machine translation (MT), and consider how to teach LLMs a new language like humans.</p><div><hr></div><h2>Low-resource MT</h2><p>To set the scene, let&#8217;s first consider what it means for a language to be considered &#8220;low-resource&#8221;. As with LLMs, the performance of MT models depends on the amount of training data&#8212;both parallel and monolingual&#8212;available in a given language. As a result, there is a <a href="https://aclanthology.org/2020.acl-main.560/">gulf between languages with lots of data and languages with little data</a> in common pre-training corpora. The latter are typically referred to as &#8220;low-resource&#8221;.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>To bridge the gap between resource-rich and resource-poor languages and make machine translation more accessible, new translation benchmarks have been created that cater specifically to low-resource languages. The <a href="https://www2.statmt.org/wmt23/">Conference on Machine Translation (WMT)</a> now regularly hosts shared tasks on low-resource MT such as for <a href="https://www2.statmt.org/wmt23/indic-mt-task.html">Indic</a> and <a href="https://www2.statmt.org/wmt23/african-mt-task.html">African languages</a>; workshops such as <a href="https://turing.iimas.unam.mx/americasnlp/2023_st.html">AmericasNLP</a> support <a href="https://aclanthology.org/2023.americasnlp-1.23/">indigenous languages</a>; and large-scale decentralized collaborations such as <a href="https://www.masakhane.io/">Masakhane</a>, <a href="https://github.com/SEACrowd/seacrowd-datahub">SEACrowd</a> and <a href="https://cohere.com/research/aya">Aya</a> created MT datasets for <a href="https://aclanthology.org/2020.findings-emnlp.195/">African languages</a>, <a href="https://aclanthology.org/2023.eacl-main.57/">Indonesian languages</a>, and <a href="https://arxiv.org/abs/2402.06619">100+ languages</a> respectively. Recently, <a href="https://github.com/facebookresearch/flores/blob/main/flores200/README.md">FLORES-200</a> expands translation data coverage to 200 languages. Beyond theses efforts, through extensive work on data cleaning, filtering, and language identification, researchers have been able to obtain data and train MT models for 1000+ languages (<a href="https://arxiv.org/abs/2205.03983">Bapna et al., 2022</a>; <a href="https://arxiv.org/abs/2207.04672">NLLB Team, 2022</a>).</p><p>LLMs are typically trained on parallel data (<a href="https://aclanthology.org/2021.acl-short.87/">Kale et al., 2021</a>) and are increasingly used for translation (<a href="https://aclanthology.org/2023.acl-long.859/">Vilar et al., 2023</a>). However, in light of ever-larger pre-training datasets, the opacity of pre-training data, and challenges of language identification on the web (<a href="https://aclanthology.org/2020.coling-main.579/">Caswell et al., 2020</a>), it is unclear how much data LLMs have seen in a low-resource language during pre-training. Chances are that most LLMs have seen <em>some</em> data in most languages that are available on the web.</p><p>From an experimental perspective, it is thus not straightforward to assess how much data LLMs <em>actually</em> need to learn to translate in a new language. While we could restrict the languages a model is trained on and examine its performance on a held-out language with increasing amounts of training data, we know that most LLM capabilities only emerge at scale. So how can we study the translation abilities of fully pre-trained LLMs in a controlled setting?</p><div><hr></div><h2>True Zero-shot MT</h2><p>To study this setting, we have to look at translating a language that was <em>truly unseen</em> during pre-training.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> For simplicity, I use the term <strong>true zero-shot MT</strong> to designate the setting of translating to a language with no pre-training data using only in-context learning data.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>While we can find data for 1500+ languages on the web (<a href="https://arxiv.org/abs/2205.03983">Bapna et al., 2022</a>), there are around 7000+ languages spoken around the world. So the remaining 5500 languages with little or no presence online are potential candidates for our target language.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>As expected, for these languages there is very little data available that has been used in standard LLM pipelines. Interestingly, the resources that <em>are</em> available for these languages are similar to those humans might use to learn a <a href="https://en.wikipedia.org/wiki/Second_language">second language (L2)</a> including:<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><ol><li><p>a list of words and their translations to learn a language&#8217;s vocabulary;</p></li><li><p>paired sentences to learn about word usage, word order, and some morphology;</p></li><li><p>a grammar book to study the structure of the language.</p></li></ol><p>Let&#8217;s take a look at how these resources can be used by LLMs:</p><h3>1. A Bilingual Word List</h3><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!36sO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd92e197c-ae1e-4709-b629-1117b2448496_2204x378.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!36sO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd92e197c-ae1e-4709-b629-1117b2448496_2204x378.png 424w, https://substackcdn.com/image/fetch/$s_!36sO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd92e197c-ae1e-4709-b629-1117b2448496_2204x378.png 848w, https://substackcdn.com/image/fetch/$s_!36sO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd92e197c-ae1e-4709-b629-1117b2448496_2204x378.png 1272w, https://substackcdn.com/image/fetch/$s_!36sO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd92e197c-ae1e-4709-b629-1117b2448496_2204x378.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!36sO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd92e197c-ae1e-4709-b629-1117b2448496_2204x378.png" width="1456" height="250" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d92e197c-ae1e-4709-b629-1117b2448496_2204x378.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:250,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:72438,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!36sO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd92e197c-ae1e-4709-b629-1117b2448496_2204x378.png 424w, https://substackcdn.com/image/fetch/$s_!36sO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd92e197c-ae1e-4709-b629-1117b2448496_2204x378.png 848w, https://substackcdn.com/image/fetch/$s_!36sO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd92e197c-ae1e-4709-b629-1117b2448496_2204x378.png 1272w, https://substackcdn.com/image/fetch/$s_!36sO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd92e197c-ae1e-4709-b629-1117b2448496_2204x378.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Database coverage of bilingual word lists in <a href="https://panlex.org/">PanLex</a>.</figcaption></figure></div><p>While lists of words and their translation only teach us a limited amount about a new language, they are available for a huge number of languages thanks to projects such as <a href="https://panlex.org/">PanLex</a>. However, for many languages, they only contain translation pairs for a small number of terms such as the numbers from 1&#8211;10.</p><p>Bilingual lexicons have been a core resource for aligning word embeddings across languages (<a href="https://arxiv.org/abs/1706.04902">Ruder et al., 2019</a>). They also have been used to bootstrap unsupervised MT systems (<a href="https://arxiv.org/abs/1710.11041">Artetxe et al., 2018</a>), though, translating text word-by-word does not get you far, even if you account for word order differences (<a href="https://arxiv.org/abs/1711.00043">Lample et al., 2018</a>).</p><p>On the other hand, even in the best LLMs, cross-lingual alignment of words in the vocabulary is still not perfect. Bilingual lexicons are thus a potential viable&#8212;and large-coverage&#8212;form of supervision. There are multiple ways in which they can be used with LLMs:</p><ul><li><p><strong>data augmentation</strong>: &#8220;noisy&#8221; code-mixed samples are created by replacing words in source language sentences with their target language translations (<a href="http://Expanding Pretrained Models to Thousands More Languages via Lexicon-based Adaptation">Wang et al., 2022</a>, <a href="https://aclanthology.org/2022.naacl-main.58/">Reid &amp; Artetxe, 2022</a>);</p></li><li><p><strong>lexical prompting</strong>: prepending <code>&lt;source word, translation&gt;</code> pairs to the prompt for source words that occur in the input (<a href="https://arxiv.org/abs/2302.07856">Ghazvininejad et al., 2023</a>);</p></li><li><p><strong>parallel data</strong>: <code>&lt;source word, translation&gt;</code> pairs are treated as &#8220;sentence&#8221; pairs and added to the existing parallel data.</p></li></ul><p><a href="https://aclanthology.org/2023.emnlp-main.26/">Jones et al. (2023)</a> compare several augmentations. They find that they provide similar gains for unsupervised machine translation and that high-quality bilingual lexicons such as <a href="https://github.com/google-research/url-nlp/tree/main/gatitos">GATITOS</a> are crucial. Recently, <a href="https://arxiv.org/abs/2402.02113">Koto et al. (2024)</a> use PanLex to extend sentiment lexicons to more languages.</p><h3>2. Few Parallel Sentences</h3><p>Parallel data is the bread and butter of MT research. MT models and LLMs are trained on millions of parallel sentences. Prior work on low-resource MT such as for Nepalese&#8211;English used 100,000s of parallel sentences (<a href="https://aclanthology.org/D19-1632/">Guzm&#225;n et al., 2019</a>).</p><p>More recent work studied the impact of around 1&#8211;6k professionally translated sentences across different low-resource languages (<a href="https://aclanthology.org/D19-1632/">Maillard et al., 2023</a>). They find that already 1,000s of high-quality parallel sentences are helpful while multilingual training and back-translation are crucial to achieve good performance. </p><p>But what can you do if you only have very few parallel sentences for a language? Such scenarios are common in the form of <em>Rosetta Stone</em> puzzles in Linguistic Olympiads.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!c_Ry!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf229425-88b5-4f0b-8755-2754f1b71671_2558x1128.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!c_Ry!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf229425-88b5-4f0b-8755-2754f1b71671_2558x1128.png 424w, https://substackcdn.com/image/fetch/$s_!c_Ry!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf229425-88b5-4f0b-8755-2754f1b71671_2558x1128.png 848w, https://substackcdn.com/image/fetch/$s_!c_Ry!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf229425-88b5-4f0b-8755-2754f1b71671_2558x1128.png 1272w, https://substackcdn.com/image/fetch/$s_!c_Ry!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf229425-88b5-4f0b-8755-2754f1b71671_2558x1128.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!c_Ry!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf229425-88b5-4f0b-8755-2754f1b71671_2558x1128.png" width="1456" height="642" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cf229425-88b5-4f0b-8755-2754f1b71671_2558x1128.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:642,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:460605,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!c_Ry!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf229425-88b5-4f0b-8755-2754f1b71671_2558x1128.png 424w, https://substackcdn.com/image/fetch/$s_!c_Ry!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf229425-88b5-4f0b-8755-2754f1b71671_2558x1128.png 848w, https://substackcdn.com/image/fetch/$s_!c_Ry!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf229425-88b5-4f0b-8755-2754f1b71671_2558x1128.png 1272w, https://substackcdn.com/image/fetch/$s_!c_Ry!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf229425-88b5-4f0b-8755-2754f1b71671_2558x1128.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">An example <em>Rosetta Stone</em> puzzle <a href="https://aclanthology.org/2020.acl-main.115/">(Sahin et al., 2020)</a>.</figcaption></figure></div><p>These puzzles are designed so that only a minimum number of parallel sentences are required to deduce the relevant translations to and from the target language (Chickasaw above). <a href="https://aclanthology.org/2020.acl-main.115/">Sahin et al. (2020)</a> collected 100 such puzzles. LLMs such as GPT-3.5 Turbo still fall short of solving these puzzles with an Exact Match score of around 20% and even advanced prompting methods are not helpful (<a href="https://aclanthology.org/2023.rocling-1.33/">Lin et al., 2023</a>).</p><h3>3. A Grammar Book</h3><p>While both bilingual lexicons and parallel sentences are well-known data sources for training MT models, grammar books are much less common. A reference grammar is a key product of linguistic fieldwork, which can also result in other artifacts such as language learning materials including alphabet books, learners&#8217; guides, etc.</p><p>Many reference grammars are available as books in pdf format. As part of sourcing open-source OCR data from Google Books for <a href="https://github.com/google-research/xtreme-up">XTREME-UP</a> (<a href="https://aclanthology.org/2023.findings-emnlp.125/">Ruder et al., 2023</a>), we came across many public-domain books in under-represented languages that described the grammar or vocabulary of the language and were created as part of missionary efforts and linguistic fieldwork.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p><p>Compared to these, <em>A grammar of Kamalang</em> (<a href="https://langsci-press.org/catalog/book/344">Visser, 2022</a>) used in MTOB is a more recent reference grammar. Reference grammars are <a href="https://guides.tricolib.brynmawr.edu/c.php?g=285265&amp;p=1900305">publicly available</a> in <a href="https://linguistic-typology.org/grammarwatch/all/by-macro-area/">100s of languages</a> and present a promising and&#8212;with exception of MTOB&#8212;untapped resource for studying the language acquisition of LLMs.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tQ8h!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tQ8h!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png 424w, https://substackcdn.com/image/fetch/$s_!tQ8h!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png 848w, https://substackcdn.com/image/fetch/$s_!tQ8h!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png 1272w, https://substackcdn.com/image/fetch/$s_!tQ8h!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tQ8h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png" width="1456" height="680" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:680,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:248952,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tQ8h!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png 424w, https://substackcdn.com/image/fetch/$s_!tQ8h!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png 848w, https://substackcdn.com/image/fetch/$s_!tQ8h!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png 1272w, https://substackcdn.com/image/fetch/$s_!tQ8h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ffbde3-d3fc-49b9-a596-ba4ca2c0ae4c_1958x914.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">An excerpt from <em>A grammar of Kamalang</em>.</figcaption></figure></div><h3>4. Putting Everything Together: MTOB</h3><p>MTOB (Machine Translation from One Book; <a href="https://arxiv.org/abs/2309.16575">Tanzer et al., 2023</a>) is a recent dataset that provides the three above resources for <a href="https://en.wikipedia.org/wiki/Kalamang_language">Kalamang</a>, an endangered language spoken by less than 200 people<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>, which is essentially absent from pre-training corpora.</p><p>In addition, it is not closely related to other languages with many speakers (which is important to measure true zero-shot performance) and uses the Latin script, which makes it easy to process with LLMs. The authors obtained the permission of the Kalamang-speaking community for using their data, which is crucial for this type of work.</p><p>Compared to standard MT evaluation where native speaker translations are the benchmark to beat, producing native-level translations may not be feasible using only the provided resources. The authors thus provide another human baseline: The first author Garrett Tanzer learned how to translate Kalamang from scratch by reading the grammar book for 10+ hours and then used the parallel data and the Internet as reference when performing the translation task over the course of several weeks. <em>That</em> is dedication to human evaluation!</p><p>In their evaluation, all LLMs underperform the human baseline, with Claude 2 performing best in comparison. Recently, <a href="https://storage.googleapis.com/deepmind-media/gemini/gemini_v1_5_report.pdf">Gemini 1.5 Pro</a> using its extremely long context window is able to use in-context learning on the Kalamang resources to improve substantially on the English-&gt;Kalamang translation task. Let&#8217;s look at what these results mean for the field overall and for under-represented languages in particular.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">NLP News is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>Future Impact</h2><h4>Long Context Modeling</h4><p>It&#8217;s quite remarkable how long-context models have progressed over the last years. Being able to translate into a new language using only in-context learning based on existing linguistic resources is an impressive feat. Nevertheless, parameter-efficient fine-tuning and retrieval-augmented generation (RAG) are important baselines for this setting that can put the in-context learning performance in perspective.</p><h4>Long Context Datasets</h4><p>While many challenging long-context benchmarks are synthetic in nature, including the most challenging tasks in Long Range Arena (<a href="https://openreview.net/forum?id=qVyeW-grC2k">Tay et al., 2021</a>) and the needle-in-the-haystack tasks in the Gemini 1.5 evaluation, datasets like MTOB provide a more realistic setting using standard linguistic resources. In addition, they enable comparison to a human baseline and to a (potentially unattainable) human native speaker. Given the increasing popularity of long-context models, I hope we see more long-context datasets grounded in realistic data and human-level comparisons.</p><h4>Under-represented Languages</h4><p>While these results highlight the potential of LLMs to learn to translate with very little data, many under-represented languages are primarily spoken; they do not have a written tradition or a standardized orthography. Text-based NLP technology is thus of limited use to them. For these languages, multi-modal LLMs will be an important foundation. Given the powerful capabilities of LLMs, we should design evaluations with the needs of the language communities in mind. A promising area is to support &#8216;contact languages&#8217; including creoles and regional varieties of standardized languages (<a href="https://aclanthology.org/2022.acl-long.539/">Bird, 2022</a>).</p><h4>NLP and Cognitive Science</h4><p>In order to study language acquisition in language models, benchmarks such as the <a href="https://babylm.github.io/">BabyLM Challenge</a> employ a developmentally plausible corpus including mostly transcribed speech and a limited number of tokens. However, the embodied, interactive, and multi-modal nature of <a href="https://en.wikipedia.org/wiki/First_language">first language (L1)</a> acquisition is challenging to replicate with current models. L2 acquisition in the form of true zero-shot MT may be a more accessible testbed to study how a model learns a new language based on limited linguistic resources.</p><h4>Interpretability and Model Understanding</h4><p>Grammar books and language learning resources similarly provide a means to analyze how LLMs acquire new information and how they use their existing knowledge to reason over new inputs. They can be used to better understand models&#8217; inner workings via tasks such as grammar induction (<a href="https://openreview.net/forum?id=H1xPR3NtPB">Kim et al., 2020</a>). For instance, an interesting question is whether the sentences a model retrieves via RAG are similar to those consulted by a human when translating into a new language.</p><h4>NLP and Linguistics</h4><p>To make most use of such resources, both in obtaining and understanding the data as well as interpreting model results requires collaborating with linguists. MTOB provides an example of how such a collaboration can look like in practice, with the linguist actively participating in the research and co-authoring the paper. Such inter-disciplinary collaborations, while challenging and complex, are often a breath of fresh air&#8212;so I hope to see more of them in the future.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Note that the term &#8220;low-resource&#8221; is typically a misnomer as many low-resource languages have data available; it may just not be easily accessible due to being in different formats, in another modality, etc. I&#8217;ll still use it here as we will strictly talk about the data available in pre-training for these languages.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>We have studied transfer to languages with unseen scripts in prior work (<a href="http://For simplicity, I use the term true zero-shot MT to designate the setting of translating to a language with no pre-training data using only in-context learning data.">Pfeiffer et al., 2021</a>) though many of these scripts have likely been seen by more recent LLMs.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>I use the term <strong>true zero-shot MT</strong> to contrast with <strong>zero-shot MT</strong> (<a href="https://arxiv.org/abs/1611.04558">Johnson et al., 2016</a>), which refers to the setting where a language has monolingual but no parallel data in pre-training. The term also relates to the usage of the term <strong>true few-shot learning</strong> (<a href="https://arxiv.org/abs/2105.11447">Perez et al., 2021</a>), which does not use any other held-out examples.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>There are many factors that affect the choice of the target language for true zero-shot MT: Many languages don&#8217;t have a standardized orthography; others use a different script, which may complicate processing; connecting with native speakers and finding data may be difficult for many languages; language similarity plays an important role for learning a new language; finally, some communities do not approve the use of MT for their language.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Note that this is a text-centric perspective of zero-shot MT. In practice, for such local languages we can also expect the existence of raw speech with translations (<a href="https://aclanthology.org/2022.acl-long.539/">Bird, 2022</a>).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>We excluded these books from our data as they are mainly in English with examples in the target language rather than the monolingual target language data that we wanted to collect.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Kalamang is spoken in the Western part of the island of New Guinea&#8212;part of Indonesia&#8212;while the Eastern part of the island is part of Papua New Guinea. Incredibly, Papua New Guinea and Indonesia are also the two countries with the <a href="https://en.wikipedia.org/wiki/Number_of_languages_by_country">most languages in the world</a>. See (<a href="https://aclanthology.org/2022.acl-long.500/">Koto et al., 2022</a>) for an overview of NLP challenges for Indonesian languages and this <a href="https://aclanthology.org/2023.ijcnlp-tutorials.2/">AACL 2023 Tutorial</a> for the current status of NLP in Southeast Asia. </p></div></div>]]></content:encoded></item><item><title><![CDATA[Thoughts on the 2024 AI Job Market]]></title><description><![CDATA[And Why I Joined Cohere]]></description><link>https://newsletter.ruder.io/p/thoughts-on-the-2024-ai-job-market</link><guid isPermaLink="false">https://newsletter.ruder.io/p/thoughts-on-the-2024-ai-job-market</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Mon, 12 Feb 2024 09:51:17 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/ac597596-ba0b-4a7b-8aaa-6b04118d690e_1400x814.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It&#8217;s crazy how the AI and NLP landscape has evolved over the last five years. 5 years ago, around the time I finished my PhD, if you wanted to work on cutting-edge natural language processing (NLP), your choice was relatively limited.</p><p>Recently, I decided to go on the job market again, which has become much more diverse. In this post, I want to highlight some macro trends that I observed and the reasons that I joined my new company, <a href="https://cohere.com/">Cohere</a>, which may be helpful in guiding your own job search.</p><p><em>Note: This post reflects my personal opinions and not those of my previous and current employers. It is written from my perspective as a Europe-based researcher focused on NLP. If you are interested in AI companies but have a different skillset, some of these thoughts should still be relevant to you.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">NLP News is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>AI Job Market Trends</h2><h3>1. Research has become more applied.</h3><p>In the past, most problems at the forefront of ML and NLP were firmly in the purview of fundamental or basic research. As models were not powerful enough, datasets reflected simplified evaluation settings that were feasible at the time and typically far removed from applications. In order to work on such cutting-edge problems, you generally had the choice of joining academia or going to a handful of big tech labs (Google Research/Brain, DeepMind, FAIR, MSR, etc).</p><p>For research advances to make their way into products could take a team months or years of dedicated work&#8212;if it succeeded at all. An exception is machine translation where research breakthroughs such as the emergence of <a href="https://blog.research.google/2006/04/statistical-machine-translation-live.html">statistical MT</a> and <a href="https://blog.research.google/2016/09/a-neural-network-for-machine.html">neural MT</a> resulted in concrete product improvements.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> On the other hand, applied research departments directly worked on improving a specific application.</p><p>In light of the emergence of pre-training (<a href="https://thegradient.pub/nlp-imagenet/">NLP&#8217;s &#8216;ImageNet moment&#8217;</a>) and models becoming more powerful, the gap between fundamental and applied research in NLP has consistently narrowed: The integration of <a href="https://aclanthology.org/N19-1423/">BERT</a>-based representations led to <a href="https://blog.google/products/search/search-language-understanding-bert/">one of the biggest quality leaps</a> in Google Search history and the recent generation of large language models (LLMs) enabled a plethora of new applications.</p><p>Problems that were previously in the domain of basic research (how to measure generation quality, how to teach models to reason, how to learn with long-range dependencies, etc) now impact real-world applications. As a result, new advances in research have the potential to have a much broader impact. This leads to many new opportunities and emerging research areas. However, it also means that researchers must consider challenges regarding the safe and responsible use of such technology.</p><p>As the Generative AI space heats up, new research breakthroughs are perceived to provide an edge over the competition. As a result, publishing them has become more challenging. In addition, with increasing proximity to product, doing purely curiosity-driven research has become more difficult: when most other research has immediate product impact, how do you justify working on an unproven direction? For researchers, this requires balancing short-term impact with long-term research potential, with the scale tilting to the former.</p><p>To encourage long-term research in this application-oriented climate, companies need to create an environment that still rewards open-ended directions. With companies focusing on applications, researchers in academia should focus on unexplored directions and look to the <a href="https://en.wikipedia.org/wiki/Blue_skies_research">blue skies</a>. While compute requirements for research have increased, there are plenty of less compute-intensive directions:</p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;235a46d3-f6d1-428d-a4fd-a766512a9b66&quot;,&quot;caption&quot;:&quot;Update Dec 30: Added mentions of BabyLM and the Languini Kitchen.&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;NLP Research in the Era of LLMs&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:7965403,&quot;name&quot;:&quot;Sebastian Ruder&quot;,&quot;bio&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/17fdd4c3-a575-4fe4-b58e-d876b78bfe2f_2416x2416.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2023-12-19T09:53:36.237Z&quot;,&quot;cover_image&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://newsletter.ruder.io/p/nlp-research-in-the-era-of-llms&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:138941502,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:55,&quot;comment_count&quot;:4,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;NLP News&quot;,&quot;publication_logo_url&quot;:&quot;https://s3.amazonaws.com/revue/profiles/images/000/027/670/thumb/sebastian_ruder_profile_photo_square.jpg?1616408613&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><h3>2. Startups are a serious alternative to a PhD.</h3><p>When people have asked me for advice in the past on whether they should do a PhD, I generally told them that it&#8217;s well worth the time investment. Not only does it unlock research jobs that require a PhD, it also is a great way to focus on your personal growth.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>In light of the applied nature of current research problems, there is another path that exposes you to cutting-edge AI work: joining a startup. To be sure, startups&#8212;particularly early-stage ones&#8212;are not for everyone. They require a person with a certain type of mindset and motivation: Someone who enjoys solving real-world problems and having a tangible, direct impact; who can work autonomously and without much guidance; who thrives in a hectic, unstructured environment and can handle ambiguity.</p><p>But if you&#8217;re comfortable in these conditions, you can acquire certain skills and knowledge much faster than in a typical PhD. You will also get hands-on experience with emerging methodologies such as instruction and preference tuning, red-teaming, LLM alignment, etc, which can prove invaluable for your career. Of course, in a startup, your work is typically determined by the company&#8217;s needs rather than your own interests so you need to be flexible.</p><p>A PhD is still the best option for you if you would like to follow your own curiosity and focus on your personal development; if you enjoy to go deep and fully dedicate yourself to a topic; if you value collaboration and mentorship; and if you enjoy being creative and coming up with genuinely new ideas. A scientific mindset as well as other research-related skills such as designing ablations and testing for hypotheses, publishing, and developing <a href="https://colah.github.io/notes/taste/">research taste</a>, among many others, are also more easily learned during a PhD.</p><h3>3. ML has become less open and more polarized.</h3><p>An amazing attribute of the ML community has been that much of ML development and research has been conducted in the open. The top ML journal JMLR was <a href="https://www.jmlr.org/history.html">founded in 2000</a> with the goal to provide open access to its publications. Conferences provide free access to their proceedings online. ML frameworks such as TensorFlow and PyTorch were originally developed by companies and then open-sourced. In NLP, journals and conferences are similarly open-access and open-source libraries such as <a href="https://github.com/huggingface/transformers">Transformers</a> are common building blocks.</p><p>Early pre-trained models such as <a href="https://arxiv.org/abs/1802.05365">ELMo</a>, <a href="https://arxiv.org/abs/1801.06146">ULMFiT</a>, <a href="https://s3-us-west-2.amazonaws.com/openai-assets/research-covers/language-unsupervised/language_understanding_paper.pdf">GPT</a>, and <a href="https://arxiv.org/abs/1910.10683">T5</a> were open-sourced as this enabled wide-spread adoption. However, this landscape of radical openness has shifted. Stalwarts of open-source AI such as OpenAI and Google have gradually released less information about their models. Starting with the first generation of LLMs, models such as <a href="https://arxiv.org/abs/2005.14165">GPT-3</a> and <a href="https://arxiv.org/abs/2204.02311">PaLM</a> were increasingly locked behind APIs&#8212;but papers still described the architecture and data in detail. More recent models such as <a href="https://arxiv.org/abs/2303.08774">GPT-4</a>, <a href="https://arxiv.org/abs/2305.10403">PaLM-2</a>, and <a href="https://arxiv.org/abs/2312.11805">Gemini</a> are not only closed-source but the corresponding papers reveal nothing about the architecture and training data.</p><p>This lack of knowledge sharing may impede progress in AI development. Fortunately, other companies and organizations continued to release a <a href="https://huggingface.co/blog/2023-in-llms">steady stream of open-source LLMs</a>. Still, even among open-source models there is a spectrum of openness. For instance, the exact composition of training data often remains a secret. Few models such as <a href="https://bigscience.huggingface.co/blog/bloom">BLOOM</a> or the recently released <a href="https://allenai.org/olmo">OLMo</a> are truly open. Among the big tech companies, <a href="https://www.theverge.com/2024/1/18/24042354/mark-zuckerberg-meta-agi-reorg-interview">Facebook</a> and <a href="https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/">Microsoft</a> showed a renewed commitment to open-source. Even Apple&#8212;with its reputation of being secretive&#8212;has been <a href="https://www.thestack.technology/apple-quietly-open-sources-key-ai-tools/">quietly open-sourcing AI projects</a>.</p><p>For industry researchers, the trend towards closed-source means that it has become harder to publish. In the past, researchers at top AI industry labs were often able to publish a steady stream of publications similar to their academic counterparts. For LLM-related papers, this stream is reduced to a trickle and new advances may eventually see the light of day only as patents rather than research publications. Additionally, it has become more difficult to publish individual contributions as advances are more likely to be absorbed into large collaborations producing a single report.</p><h3>4.  Research is concentrated in large projects.</h3><p>The average number of authors on a publication has <a href="https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0149504">steadily increased</a>. Starting in particle physics, <a href="https://www.nature.com/articles/d41586-019-03862-0">author numbers have surged</a> in recent years due to massive global collaborations such as the <a href="https://www.nature.com/articles/nature.2015.17567">Large Hadron Collider</a>. The emergence of LLMs brought this trend to ML and NLP. Recent examples of such large-scale collaborations are <a href="https://arxiv.org/abs/2211.05100">BLOOM</a> (300+ authors), <a href="https://arxiv.org/abs/2303.08774">GPT-4</a> (200+ authors), and <a href="https://arxiv.org/abs/2312.11805">Gemini</a> (900+ authors). While several successful LLMs have been produced by small teams, the number of authors of an LLM has generally increased with the number of its parameters.</p><p>LLM projects not only require people with research skills but also strong software engineers that can design systems that scale efficiently to 100s of billions of parameters and trillions of tokens. In addition, LLMs require disparate sets of expertise including data processing, optimization, fine-tuning, RL, evaluation, safety, infrastructure, multi-modality, etc. As a result and due to their strategic importance, the size of teams working on the latest generation of LLMs has rapidly increased.</p><p>This size contrasts with the previous generation of AI breakthroughs such as <a href="https://www.nature.com/articles/nature16961">AlphaGo</a>, which were executed by much smaller, focused teams. Such size poses challenges for the effective execution and prioritization, increasing friction and making it more difficult to quickly make decisions. A less direct downside of the increasing number of people getting absorbed into LLM-related research is that other research directions that do not directly relate to the latest generation of LLMs such as the development of Transformer alternatives are deprioritized.</p><h3>5.  More companies, more opportunities.</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HK5D!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48afa389-0cc3-4afb-975a-32899d16cabf_1400x814.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HK5D!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48afa389-0cc3-4afb-975a-32899d16cabf_1400x814.png 424w, https://substackcdn.com/image/fetch/$s_!HK5D!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48afa389-0cc3-4afb-975a-32899d16cabf_1400x814.png 848w, https://substackcdn.com/image/fetch/$s_!HK5D!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48afa389-0cc3-4afb-975a-32899d16cabf_1400x814.png 1272w, https://substackcdn.com/image/fetch/$s_!HK5D!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48afa389-0cc3-4afb-975a-32899d16cabf_1400x814.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HK5D!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48afa389-0cc3-4afb-975a-32899d16cabf_1400x814.png" width="1400" height="814" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/48afa389-0cc3-4afb-975a-32899d16cabf_1400x814.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:814,&quot;width&quot;:1400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:90120,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HK5D!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48afa389-0cc3-4afb-975a-32899d16cabf_1400x814.png 424w, https://substackcdn.com/image/fetch/$s_!HK5D!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48afa389-0cc3-4afb-975a-32899d16cabf_1400x814.png 848w, https://substackcdn.com/image/fetch/$s_!HK5D!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48afa389-0cc3-4afb-975a-32899d16cabf_1400x814.png 1272w, https://substackcdn.com/image/fetch/$s_!HK5D!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48afa389-0cc3-4afb-975a-32899d16cabf_1400x814.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Generative AI startups (Credit: <a href="https://medium.com/@dawncapitalteam/not-all-hype-generative-ai-stars-emerging-in-europe-453b335907b4">Dawn Capital</a>).</figcaption></figure></div><p>The advent of LLMs led to a wave of new companies leveraging this technology&#8212;and prompted existing companies to figure out how incorporate these models into their products. YC, the prolific startup incubator, has already funded more than <a href="https://www.ycombinator.com/companies/industry/generative-ai">100 generative AI startups</a>. A recent <a href="https://www.mckinsey.com/capabilities/mckinsey-digital/our-insights/the-economic-potential-of-generative-ai-the-next-productivity-frontier#key-insights">McKinsey report</a> estimated that Generative AI&#8217;s impact on productivity could add trillions of dollars to the global economy, with most of the expected value concentrated across four use cases: customer operations, marketing and sales, software engineering, and R&amp;D.</p><p>Generative AI, however, is still just at the beginning. Many research challenges remain including mitigating hallucinations and ensuring trustworthiness and attribution, aligning models to reliably elicit desired behavior, ensuring robust reasoning, etc. In order to effectively use LLMs for business use cases, we furthermore need to <a href="https://www.forbes.com/sites/peterbendorsamuel/2024/01/08/reasons-why-generative-ai-pilots-fail-to-move-into-production/?sh=551933d36b4a">successfully conduct pilot studies</a>, assess biases and risks, define suitable guardrails, rethink core business processes, and develop new skills in the workforce, among others.</p><p>With all of these new AI companies, it is difficult to choose the one that is the best fit for you.</p><div><hr></div><h2>Why I Joined Cohere</h2><p>Below, I highlight the criteria that led me to join Cohere. Many of these considerations are personal but I hope they may be useful to you as inspiration or to guide your own job search.</p><h3><strong>1.  Openness and community</strong></h3><p>In addition to building powerful proprietary enterprise models, Cohere supports openness and inclusion through its non-profit research arm, Cohere for AI (C4AI). C4AI&#8217;s openness is not an after-thought but part of Cohere&#8217;s DNA: the idea for C4AI goes back to FOR.ai, a decentralized ML collaboration&#8212;which I highlighted in a <a href="https://newsletter.ruder.io/p/nlp-progress-restrospectives-and-look-ahead-new-nlp-courses-independent-research-initiatives-interviews-lots-of-resources-217744?utm_source=%2Fsearch%2Ffor.ai&amp;utm_medium=reader2">2020 newsletter</a>&#8212;initiated by Cohere founders Aidan Gomez and Ivan Zhang, among others. C4AI <a href="https://txt.cohere.com/c4ai-2023/">published more than 30 papers in 2023</a>. Talented researchers from a diversity of backgrounds are mentored by researchers across the organization in their <a href="https://txt.cohere.com/c4ai-scholars-program/">Scholars Program</a>. C4AI organizes large-scale community initiatives including <a href="https://txt.cohere.com/aya-multilingual/">Aya</a>, a massive collaboration to develop a large multilingual open-source instruction tuning dataset and model. In addition, Cohere also invests in programs that make ML more accessible such as <a href="https://docs.cohere.com/docs/llmu">LLM University</a>.</p><h3>2.  A mature start-up</h3><p>At an early-stage startup, you can move fast but things are hectic and unstructured. In big tech, there are established tools and processes available but bureaucracy can impede progress.</p><p>Cohere occupies a middle ground. It has been around for a while and had time to build structure and processes, which make it easy to hit the ground running and directly have an impact without getting bogged down by unrelated tasks. The core components of the LLM pipeline are firmly in place and are being refined and iterated upon.</p><p>At the same time, the team is small enough so that you can have impact across the entire LLM stack and own crucial parts of the pipeline. There is little friction and red tape, which makes it easy to prototype and test new improvements and to collaborate across teams.</p><h3>3.  Enabling remote work</h3><p>With a new baby, having a flexible working arrangement that would allow me to work remotely was very important to me. However, not every company that allows its employees to work from home is set up for remote work. It&#8217;s good to be aware: Does the culture support remote work or will you miss out on conversations at the micro-kitchen? Do the tools enable working remotely? Are there opportunities to meet in person?</p><p>For companies with multiple offices, an important factor is where the company&#8217;s main office or a project&#8217;s center of gravity is located as it may be harder to achieve the same level of impact if you are in a satellite office or working remotely. In the same vein, working with colleagues in similar timezones often facilitates collaboration. At the beginning of your career, it is often a good idea to work in-person as it will be easier to seek advice and learn from colleagues.</p><p>At Cohere, half of all employees are remote. Tools and benefits support remote work and events provide equal access for distributed teams. Cohere has offices in Toronto, London, San Francisco, and Palo Alto that you can visit to work with team members in-person. Many people working on ML and modeling at Cohere are based in Europe or in the US East Coast so it&#8217;s easy to collaborate across these timezones. </p><h3>4.  Alignment</h3><p>Another thing that was important to me is to be aligned with the overall goals of the company. While most companies are profit-driven, some are more serious about having a positive impact on society than others. This is reflected in the way they develop their technology (do they prioritize safety and ethics and put appropriate safeguards in place?) as well as the programs they organize, causes they support, and the way they interact with the community.</p><p>Cohere aims to develop AI models in a responsible manner to serve humanity, an important objective. Making models more accessible across languages is not only a personal objective but serves this overarching goal as it enables Cohere&#8217;s customers to reach their users across the world.</p><h3>5.  Team and culture</h3><p>Cohere&#8217;s team is world-class. I had worked with a few at Google DeepMind and knew others from conferences or by reputation. Cohere has teams with deep expertise across many LLM areas such as pre-training, RL, retrieval augmentation and search. The latter are crucial for knowledge-intensive and enterprise use cases. Beyond research, senior leaders at Cohere have experience in building and scaling products and businesses to billions of users.</p><p>In addition, I enjoyed all my interactions and conversations with Cohere employees. Everyone I got to know was kind, humble, and genuine. The culture is collaborative and everyone is aligned towards the same objective.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> People are motivated by helping each other succeed.</p><p>For more information about careers at Cohere, check out the <a href="https://cohere.com/careers">careers page</a>. Cohere is <a href="https://jobs.lever.co/cohere/">hiring across many roles</a>.</p><p><em>For a job market perspective from an RLHF researcher, check out </em><span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Nathan Lambert&quot;,&quot;id&quot;:10472909,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fdda47b96-836a-4b95-99a0-f0ec744d4245_2316x2316.jpeg&quot;,&quot;uuid&quot;:&quot;ce7e82ae-b9a7-47bd-8f07-381fd5437ae5&quot;}" data-component-name="MentionToDOM"></span> &#8216;s <a href="https://www.interconnects.ai/p/ai-research-job-market">Interconnects post</a>.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Other examples include OCR, image recognition, <a href="https://en.wikipedia.org/wiki/Predictive_text">predictive text</a>, <a href="https://blog.research.google/2016/05/announcing-syntaxnet-worlds-most.html">dependency parsing</a> (when it was still widely used as an intermediate step in NLP pipelines), among others.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>I&#8217;ve collected some advice in <a href="https://www.ruder.io/10-tips-for-research-and-a-phd/">10 tips for research and a PhD</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>A company&#8217;s culture is difficult to assess from the outside. Try to ask current and former employees about their work environment and read reviews of the company online. </p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[The Big Picture of AI Research]]></title><description><![CDATA[A Workshop Retrospective]]></description><link>https://newsletter.ruder.io/p/the-big-picture-of-ai-research</link><guid isPermaLink="false">https://newsletter.ruder.io/p/the-big-picture-of-ai-research</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Thu, 18 Jan 2024 07:09:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!nIVS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>More papers on AI are published than ever before<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> but each paper tends to only present its part of the picture&#8212;and it becomes difficult to recognize the larger story to which a paper is connected. To encourage the community to explore broader research narratives, we<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> co-organized the <a href="https://www.bigpictureworkshop.com/">Big Picture Workshop</a> at <a href="https://2023.emnlp.org/">EMNLP 2023</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> We received a number of <a href="https://www.bigpictureworkshop.com/">high-quality submissions</a> that distill important research topics, from narrative understanding to modern generation techniques.</p><p>My favorite part of the workshop, however, were the invited talks. We had asked researchers from different labs working on the same topic to reflect on and consolidate their often disagreeing contributions. The result were talks that were more nuanced and engaging than the&#8212;typically one-sided&#8212;scientific presentations. The talks covered topics from in-context learning to attention as explanation and morality.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nIVS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nIVS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png 424w, https://substackcdn.com/image/fetch/$s_!nIVS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png 848w, https://substackcdn.com/image/fetch/$s_!nIVS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png 1272w, https://substackcdn.com/image/fetch/$s_!nIVS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nIVS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png" width="1456" height="695" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:695,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2195068,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nIVS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png 424w, https://substackcdn.com/image/fetch/$s_!nIVS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png 848w, https://substackcdn.com/image/fetch/$s_!nIVS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png 1272w, https://substackcdn.com/image/fetch/$s_!nIVS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4f7ffd8-aa45-40f5-b4f5-c962b89fca2a_1676x800.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The audience at the Big Picture Workshop.</figcaption></figure></div><div><hr></div><p><em>I recently joined Cohere to help solve real-world problems with LLMs. <a href="https://cohere.com/careers">We are hiring!</a> </em></p><div><hr></div><h2>What Does In-Context Learning Need?</h2><p>In-context learning (ICL) is one of the most important emerging phenomena of LLMs but it is still not clearly understood what factors contribute to its success. At EMNLP 2022, two papers with seemingly contradictory hypotheses were published: In &#8220;<a href="https://aclanthology.org/2022.emnlp-main.759.pdf">Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?</a>&#8221;, <a href="https://shmsw25.github.io/">Sewon Min</a> and others argued that random labels perform similarly to using ground-truth labels in the demonstrations. On the other hand, <a href="https://scholar.google.com/citations?user=BqaWtH8AAAAJ&amp;hl=en">Kang Min Yoo</a> and <a href="http://ids.snu.ac.kr/site/members/M_Junyeob_Kim.html">Junyeob Kim</a> found that ground-truth labels <em>are</em> important in &#8220;<a href="https://aclanthology.org/2022.emnlp-main.155.pdf">Ground-Truth Labels Matter: A Deeper Look into Input-Label Demonstrations</a>&#8221;. So, do we need ground-truth labels for ICL or not?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GRgi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ad2fd2b-9a4b-408c-a4f6-f220c93afacb_1940x830.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GRgi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ad2fd2b-9a4b-408c-a4f6-f220c93afacb_1940x830.png 424w, https://substackcdn.com/image/fetch/$s_!GRgi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ad2fd2b-9a4b-408c-a4f6-f220c93afacb_1940x830.png 848w, https://substackcdn.com/image/fetch/$s_!GRgi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ad2fd2b-9a4b-408c-a4f6-f220c93afacb_1940x830.png 1272w, https://substackcdn.com/image/fetch/$s_!GRgi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ad2fd2b-9a4b-408c-a4f6-f220c93afacb_1940x830.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GRgi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ad2fd2b-9a4b-408c-a4f6-f220c93afacb_1940x830.png" width="1456" height="623" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2ad2fd2b-9a4b-408c-a4f6-f220c93afacb_1940x830.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:623,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:479214,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GRgi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ad2fd2b-9a4b-408c-a4f6-f220c93afacb_1940x830.png 424w, https://substackcdn.com/image/fetch/$s_!GRgi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ad2fd2b-9a4b-408c-a4f6-f220c93afacb_1940x830.png 848w, https://substackcdn.com/image/fetch/$s_!GRgi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ad2fd2b-9a4b-408c-a4f6-f220c93afacb_1940x830.png 1272w, https://substackcdn.com/image/fetch/$s_!GRgi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ad2fd2b-9a4b-408c-a4f6-f220c93afacb_1940x830.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In their talk (<a href="https://drive.google.com/file/d/16_VrwrVz5UG-SoYgBYG4nWRZewriz1HX/view">slides</a>), Sewon and Junyeob shed light on this conundrum. ICL with random labels is more sensitive than using ground-truth labels&#8212;but still works with careful prompting. However, small deteriorations in performance can be observed, which are not negligible in real-world applications. Nevertheless, although their impact may vary across setups, the correctness of labels is still one of the core components of successful ICL.</p><p>They argue that the main reason why ICL without ground-truth labels works is because ICL activates priors from pre-training rather than learning new tasks on-the-fly. What I found very insightful is that they even put their findings in the context of a more recent work, &#8220;<a href="https://arxiv.org/abs/2303.03846">Larger language models do in-context learning differently</a>&#8221;, which hypothesizes that overriding such semantic priors is an emerging ability in larger models. Upon closer inspection of that paper&#8217;s results, however, they found that no model with flipped labels performs better than random. So with standard ICL, even large models are unable to override their pre-training priors.</p><h2>Is "Attention = Explanation"?</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!krHJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b7ea75e-9a8c-466d-a47b-532fc541c135_2621x1086.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!krHJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b7ea75e-9a8c-466d-a47b-532fc541c135_2621x1086.png 424w, https://substackcdn.com/image/fetch/$s_!krHJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b7ea75e-9a8c-466d-a47b-532fc541c135_2621x1086.png 848w, https://substackcdn.com/image/fetch/$s_!krHJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b7ea75e-9a8c-466d-a47b-532fc541c135_2621x1086.png 1272w, https://substackcdn.com/image/fetch/$s_!krHJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b7ea75e-9a8c-466d-a47b-532fc541c135_2621x1086.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!krHJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b7ea75e-9a8c-466d-a47b-532fc541c135_2621x1086.png" width="1456" height="603" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2b7ea75e-9a8c-466d-a47b-532fc541c135_2621x1086.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:603,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:364035,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!krHJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b7ea75e-9a8c-466d-a47b-532fc541c135_2621x1086.png 424w, https://substackcdn.com/image/fetch/$s_!krHJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b7ea75e-9a8c-466d-a47b-532fc541c135_2621x1086.png 848w, https://substackcdn.com/image/fetch/$s_!krHJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b7ea75e-9a8c-466d-a47b-532fc541c135_2621x1086.png 1272w, https://substackcdn.com/image/fetch/$s_!krHJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b7ea75e-9a8c-466d-a47b-532fc541c135_2621x1086.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In 2019, there was a series of papers memorably titled &#8220;<a href="https://aclanthology.org/N19-1357/">Attention is not Explanation</a>&#8221; and &#8220;<a href="https://aclanthology.org/D19-1002/">Attention is not not Explanation</a>&#8221; that studied whether attention is useful as faithful explanation of model predictions.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> In their joint talk (<a href="https://drive.google.com/file/d/1jcVezEjj_OXP7eLXd5xy8J37LLMpJ1A_/view">slides</a>), <a href="https://successar.github.io/">Sarthak Jain</a> and <a href="https://sarahwie.github.io/">Sarah Wiegreffe</a>, the first authors of the papers, reconciled the findings of their papers and contextualized them with regard to recent developments in the field.</p><p>So, is &#8220;attention = explanation&#8221;? Putting things into perspective, both authors highlight that attention mechanisms in LSTM networks can serve as faithful explanation under certain conditions; there is no one-size-fits-all answer. However, faithfulness evaluation is difficult due to a lack of a ground truth.</p><p>But how useful is attention to explain model predictions today? They highlighted that attention is no longer very useful for instance-level explanations but that it still matters for understanding the mechanisms underlying general-purpose Transformers beyond specific models, datasets, and tasks. Understanding attention is thus still important in this context.</p><h2>Can Machines Learn Morality?</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zwAS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7753ab-fbe2-4267-a0b1-448d7d027d26_3880x2332.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zwAS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7753ab-fbe2-4267-a0b1-448d7d027d26_3880x2332.jpeg 424w, https://substackcdn.com/image/fetch/$s_!zwAS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7753ab-fbe2-4267-a0b1-448d7d027d26_3880x2332.jpeg 848w, https://substackcdn.com/image/fetch/$s_!zwAS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7753ab-fbe2-4267-a0b1-448d7d027d26_3880x2332.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!zwAS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7753ab-fbe2-4267-a0b1-448d7d027d26_3880x2332.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zwAS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7753ab-fbe2-4267-a0b1-448d7d027d26_3880x2332.jpeg" width="1456" height="875" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1a7753ab-fbe2-4267-a0b1-448d7d027d26_3880x2332.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:875,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1765344,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zwAS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7753ab-fbe2-4267-a0b1-448d7d027d26_3880x2332.jpeg 424w, https://substackcdn.com/image/fetch/$s_!zwAS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7753ab-fbe2-4267-a0b1-448d7d027d26_3880x2332.jpeg 848w, https://substackcdn.com/image/fetch/$s_!zwAS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7753ab-fbe2-4267-a0b1-448d7d027d26_3880x2332.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!zwAS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7753ab-fbe2-4267-a0b1-448d7d027d26_3880x2332.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In 2021, <a href="https://liweijiang.me/">Liwei Jiang</a> led researchers from AI in training an ML model, <a href="https://arxiv.org/abs/2110.07574">Delphi</a>, to reason about ethical judgements. To train the model, they created a new dataset, Commonsense Norm Bank, containing 1.7M examples of descriptive judgements on everyday situations. This research program was critiqued by <a href="https://twitter.com/zeeraktalat">Zeerak Talat</a> and others (&#8220;<a href="https://aclanthology.org/2022.naacl-main.56/">On the Machine Learning of Ethical Judgments from Natural Language</a>&#8221;). In 2023, Liwei and others created <a href="https://arxiv.org/abs/2309.00779">Value Kaleidoscope</a>, a new dataset to model potentially conflicting human values involved in decision-making.</p><p>In their joint talk (<a href="https://drive.google.com/file/d/1GIx653ht4vJ7g6Eny2fXDXH932WO-QIB/view">slides</a>), both discussed their research agendas and the challenges of teaching morality and ethics to AI models. They also engaged in a higher-level meta discussion on disagreements in science. They observed that science and conflict go hand in hand and found honesty and good-faith behavior key to resolve such situations.</p><h2>The Vision Thing</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xPuD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44d8e5bb-f76f-4553-9974-d523f55407f0_3894x2748.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xPuD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44d8e5bb-f76f-4553-9974-d523f55407f0_3894x2748.jpeg 424w, https://substackcdn.com/image/fetch/$s_!xPuD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44d8e5bb-f76f-4553-9974-d523f55407f0_3894x2748.jpeg 848w, https://substackcdn.com/image/fetch/$s_!xPuD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44d8e5bb-f76f-4553-9974-d523f55407f0_3894x2748.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!xPuD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44d8e5bb-f76f-4553-9974-d523f55407f0_3894x2748.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xPuD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44d8e5bb-f76f-4553-9974-d523f55407f0_3894x2748.jpeg" width="1456" height="1028" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/44d8e5bb-f76f-4553-9974-d523f55407f0_3894x2748.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1028,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3390272,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xPuD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44d8e5bb-f76f-4553-9974-d523f55407f0_3894x2748.jpeg 424w, https://substackcdn.com/image/fetch/$s_!xPuD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44d8e5bb-f76f-4553-9974-d523f55407f0_3894x2748.jpeg 848w, https://substackcdn.com/image/fetch/$s_!xPuD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44d8e5bb-f76f-4553-9974-d523f55407f0_3894x2748.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!xPuD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44d8e5bb-f76f-4553-9974-d523f55407f0_3894x2748.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Finally, Raymond J. Mooney gave an excellent invited talk (<a href="https://drive.google.com/file/d/1OAqt-1qFoHT8oNYB6NCyaqd7Skv37oik/view">slides</a>) where he discussed the importance of finding and pursuing your research passion. He reviewed his changing research vision over the last 40+ years that started with explanation-based learning, took him to bridging ML and NLP, ML for semantic parsing, and more recently grounded NLP, language and code, and language and 3D animation. It&#8217;s rare to get such a personal and inspired account on the motivations behind changes in a research vision from a luminary of the field.</p><p>For any aspiring researcher, this talk is a treasure chest full of useful and practical advice. I would highly recommend watching the <a href="https://us06web.zoom.us/rec/play/Plet9ua6wpWaNkPZXP5ioR2NGDgM5eucGW5Z7qjGdIG3FoI7NrVYwAYFnS6QevZ_67Y_q-ID7IRCUBVw.NGvDC97qxjofkCf0?canPlayFromShare=true&amp;from=share_recording_detail&amp;startTime=1701908520000&amp;componentName=rec-play&amp;originRequestUrl=https%3A%2F%2Fus06web.zoom.us%2Frec%2Fshare%2FRnBM-pPFJKaCxH_4FE0ehJyPw3ZfLxNWe_9SCkylWR40KIDW5y_bey4D_PJ8g2TC.7Dy5zKUm59N50z6y%3FstartTime%3D1701908520000">recording</a> (start: 0:07:00).</p><h2>Final Takeaways</h2><p>In summary, this was one of my favorite workshops that I&#8217;ve attended or organized. It&#8217;s a breath of fresh air when talks are more than just an oral recapitulation of a paper. At its best, research is a collaborative and sometimes argumentative conversation. Scientific publications are a culmination of this process but for various reasons<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> do not provide the full picture.</p><p>This workshop started as an experiment. We wanted to see whether we can turn scientific talks into something more akin to a debate that sheds light on a topic from different perspectives. Overall, the experiment was a success. The speakers did a stellar job presenting and contextualizing their research. The audience was engaged and speakers fielded a flood of questions for each talk while we received a lot of positive feedback on the overall format. The speakers deserve special thanks, however! They put in extra effort by having multiple meetings to prepare and sync with their counterpart. Some of them had never spoken before so the workshop also served as an opportunity to form new connections.</p><p>The Big Picture Workshop won&#8217;t return this year at an NLP conference but we hope to bring it back in 2025. Feel free to reach out to <a href="https://yanaiela.github.io/">Yanai Elazar</a> with ideas and feedback. If you enjoyed the format, you are welcome to organize a similar workshop at another venue. All materials (including the proposal, task tracker, email templates, etc) are <a href="https://www.bigpictureworkshop.com/open-workshop">available online</a>. While it does take more effort to present a topic from different perspectives, we hope future workshops and presenters will consider taking a more debate-style approach to their talks.</p><p><em>Thank you to the sponsors of the workshop, Amazon, Google, and HuggingFace! Thanks to Yanai Elazar for feedback on this post.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">NLP News is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>See the <a href="https://arxiv.org/stats/monthly_submissions">arXiv monthly submission stats</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p><a href="https://yanaiela.github.io/">Yanai Elazar</a>, <a href="https://aetting.github.io/">Allyson Ettinger</a>, <a href="https://norakassner.github.io/">Nora Kassner</a>, <a href="https://nasmith.github.io/">Noah A. Smith</a>, and I.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>The workshop was inspired by other workshops that seek to add more nuance to the research conversation such as <a href="https://ml-retrospectives.github.io/">ML Retrospectives</a> and the <a href="https://aclanthology.org/volumes/2023.insights-1/">Workshop on Insights from Negative Results</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>The recording is available <a href="https://us06web.zoom.us/rec/play/Plet9ua6wpWaNkPZXP5ioR2NGDgM5eucGW5Z7qjGdIG3FoI7NrVYwAYFnS6QevZ_67Y_q-ID7IRCUBVw.NGvDC97qxjofkCf0?canPlayFromShare=true&amp;from=share_recording_detail&amp;startTime=1701908520000&amp;componentName=rec-play&amp;originRequestUrl=https%3A%2F%2Fus06web.zoom.us%2Frec%2Fshare%2FRnBM-pPFJKaCxH_4FE0ehJyPw3ZfLxNWe_9SCkylWR40KIDW5y_bey4D_PJ8g2TC.7Dy5zKUm59N50z6y%3FstartTime%3D1701908520000">here</a>. Approximate timestamps for the talks: Raymond J. Mooney (0:07:00); Sarah &amp; Sarthak (1:53:00); Liwei &amp; Zeerak (4:25:00); Sewon &amp; Junyeob (6:55:00).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>These were accompanied by Medium posts by <a href="https://medium.com/@yuvalpinter/attention-is-not-not-explanation-dbc25b534017">Yuval Pinter</a> and <a href="https://medium.com/@byron.wallace/thoughts-on-attention-is-not-not-explanation-b7799c4c3b24">Byron Wallace</a> and more recently revisited in &#8220;<a href="https://aclanthology.org/2022.acl-long.269/">Is Attention Explanation? An Introduction to the Debate</a>&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Due to space or time limitations, to create a convincing narrative, bias or lack of awareness of the authors, etc.</p></div></div>]]></content:encoded></item><item><title><![CDATA[NLP Research in the Era of LLMs]]></title><description><![CDATA[5 Key Research Directions Without Much Compute]]></description><link>https://newsletter.ruder.io/p/nlp-research-in-the-era-of-llms</link><guid isPermaLink="false">https://newsletter.ruder.io/p/nlp-research-in-the-era-of-llms</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Tue, 19 Dec 2023 09:53:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Update Dec 30</em>: Added mentions of BabyLM and the Languini Kitchen.</p><div><hr></div><p>NLP research has undergone a paradigm shift over the last year. A range of large language models (LLMs) has validated the unreasonable effectiveness of scale<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>. Currently, the state of the art on most benchmarks is held by LLMs that are expensive to fine-tune and prohibitive to pre-train outside of a few industry labs.</p><p>In the past, a barrier to doing impactful research has often been a lack of awareness of fruitful research areas and compelling hypotheses to explore<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>. In contrast, <strong>NLP researchers today are faced with a constraint that is much harder to overcome: compute.</strong></p><p>In an era where running state-of-the-art models requires a garrison of expensive GPUs, what research is left for academics, PhD students, and newcomers to NLP without such deep pockets? Should they focus on the analysis of black-box models and niche topics ignored by LLM practitioners?</p><p>In this newsletter, I first argue why the current state of research is not as bleak&#8212;rather the opposite! I will then highlight five research directions that are important for the field and do not require much compute. I take inspiration from the following reviews of research directions in the era of LLMs:</p><ul><li><p><a href="https://arxiv.org/abs/2304.06035">Togelius &amp; Yannakakis. (Mar 2023). Choose Your Weapon: Survival Strategies for Depressed AI Academics</a></p></li><li><p><a href="https://arxiv.org/abs/2305.12544">Ignat et al. (May 2023). A PhD Student's Perspective on Research in NLP in the Era of Very Large Language Models</a></p></li><li><p><a href="https://arxiv.org/abs/2310.20633">Li et al. (Oct 2023). Defining a New NLP Playground</a></p></li><li><p><a href="https://arxiv.org/abs/2311.05020">Saphra et al. (Nov 2023). First Tragedy, then Parse: History Repeats Itself in the New Era of Large Language Models</a></p></li><li><p><a href="https://aclanthology.org/2023.emnlp-main.0.pdf">Manning (Dec 2023). Academic NLP research in the Age of LLMs: Nothing but blue skies! </a><em><a href="https://aclanthology.org/2023.emnlp-main.0.pdf#page=34">EMNLP 2023 Keynote talk</a>, <a href="https://underline.io/events/431/sessions/16480/lecture/91315-academic-nlp-research-in-the-age-of-llms-nothing-but-blue-skies">recording</a> (requires EMNLP 2023 registration)</em></p></li></ul><p>I highly recommend these for different perspectives on current LLM research and for a broader overview of research topics beyond the ones presented in this article.</p><div><hr></div><h2>A Cause for Optimism</h2><p>Research is cyclical. Computer scientist and <a href="https://www.aclweb.org/adminwiki/index.php?title=ACL_Lifetime_Achievement_Award_Recipients">ACL lifetime achievement award recipient</a> Karen Sp&#228;rck Jones <a href="https://aclanthology.org/www.mt-archive.info/Zampolli-1994-Sparck-Jones.pdf">wrote in 1994</a>:</p><blockquote><p><em>Those [&#8230;] who had been around for a long time, can see old ideas reappearing in new guises [&#8230;]. But the new costumes are better made, of better materials, as well as more becoming: so research is not so much going round in circles as ascending a spiral.</em></p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vqXl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e65b73-0428-415b-be9f-7d536e6eea69_1155x372.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vqXl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e65b73-0428-415b-be9f-7d536e6eea69_1155x372.png 424w, https://substackcdn.com/image/fetch/$s_!vqXl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e65b73-0428-415b-be9f-7d536e6eea69_1155x372.png 848w, https://substackcdn.com/image/fetch/$s_!vqXl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e65b73-0428-415b-be9f-7d536e6eea69_1155x372.png 1272w, https://substackcdn.com/image/fetch/$s_!vqXl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e65b73-0428-415b-be9f-7d536e6eea69_1155x372.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vqXl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e65b73-0428-415b-be9f-7d536e6eea69_1155x372.png" width="1155" height="372" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/91e65b73-0428-415b-be9f-7d536e6eea69_1155x372.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:372,&quot;width&quot;:1155,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:255738,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vqXl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e65b73-0428-415b-be9f-7d536e6eea69_1155x372.png 424w, https://substackcdn.com/image/fetch/$s_!vqXl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e65b73-0428-415b-be9f-7d536e6eea69_1155x372.png 848w, https://substackcdn.com/image/fetch/$s_!vqXl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e65b73-0428-415b-be9f-7d536e6eea69_1155x372.png 1272w, https://substackcdn.com/image/fetch/$s_!vqXl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F91e65b73-0428-415b-be9f-7d536e6eea69_1155x372.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In the same vein, <a href="https://arxiv.org/abs/2311.05020">Saphra et al. (2023)</a> highlight the similarities between the current era of LLMs and the Statistical Machine Translation (SMT) era where translation performance was shown to scale by training a phrase-based language model on more and more web data.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Fbaw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa703d864-1d08-476e-b449-a5b680a65afb_1846x983.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Fbaw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa703d864-1d08-476e-b449-a5b680a65afb_1846x983.png 424w, https://substackcdn.com/image/fetch/$s_!Fbaw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa703d864-1d08-476e-b449-a5b680a65afb_1846x983.png 848w, https://substackcdn.com/image/fetch/$s_!Fbaw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa703d864-1d08-476e-b449-a5b680a65afb_1846x983.png 1272w, https://substackcdn.com/image/fetch/$s_!Fbaw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa703d864-1d08-476e-b449-a5b680a65afb_1846x983.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Fbaw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa703d864-1d08-476e-b449-a5b680a65afb_1846x983.png" width="1456" height="775" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a703d864-1d08-476e-b449-a5b680a65afb_1846x983.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:775,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:292004,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Fbaw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa703d864-1d08-476e-b449-a5b680a65afb_1846x983.png 424w, https://substackcdn.com/image/fetch/$s_!Fbaw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa703d864-1d08-476e-b449-a5b680a65afb_1846x983.png 848w, https://substackcdn.com/image/fetch/$s_!Fbaw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa703d864-1d08-476e-b449-a5b680a65afb_1846x983.png 1272w, https://substackcdn.com/image/fetch/$s_!Fbaw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa703d864-1d08-476e-b449-a5b680a65afb_1846x983.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Results slide of Franz Och&#8217;s keynote talk at a 2005 MT workshop. Credit: <a href="https://arxiv.org/abs/2311.05020">Saphra et al. (2023)</a>.</figcaption></figure></div><p>More recently, we have seen the success of scale with the advent of word embeddings in 2013 and the emergence of pre-trained LMs in 2018.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> In all cases, academic research was not left in the dust but went on to make contributions that shaped the next era, from KenLM (<a href="https://aclanthology.org/W11-2123/">Heafield, 2011</a>), an efficient LM library that enabled academics to outperform industry MT systems, to the word2vec alternative GloVe (<a href="https://aclanthology.org/D14-1162/">Pennington et al., 2014</a>), to pre-trained LMs developed in non-profits and academia such as ELMo (<a href="https://aclanthology.org/N18-1202/">Peters et al., 2018</a>) and ULMFiT (<a href="https://aclanthology.org/P18-1031/">Howard &amp; Ruder, 2018</a>).</p><p><strong>The main lesson here is that while massive compute often achieves breakthrough results, its usage is often inefficient. Over time, improved hardware, new techniques, and novel insights provide opportunities for dramatic compute reduction.</strong></p><p>In his <a href="https://smerity.com/articles/2018/limited_compute.html">2018 article</a>, Stephen Merity provides two examples of this trend where the first instance of a method was exorbitantly compute-intensive while only a year later, compute costs were dramatically reduced:</p><ul><li><p>New York Times (2012): "<a href="https://www.nytimes.com/2012/06/26/technology/in-a-big-network-of-computers-evidence-of-machine-learning.html">How Many Computers to Identify a Cat? </a><strong><a href="https://www.nytimes.com/2012/06/26/technology/in-a-big-network-of-computers-evidence-of-machine-learning.html">16,000 (CPU cores)</a></strong>"<br>One year later: "<strong><a href="http://proceedings.mlr.press/v28/coates13.html">three servers each with two quad-core CPUs and four Nvidia GeForce GTX 680 GPUs</a></strong>"</p></li><li><p>Neural Architecture Search: "<strong>32,400-43,200 GPU hours</strong>"<br>Just over a year later: "<strong>single Nvidia GTX 1080Ti GPU</strong>, the search for architectures takes <strong>less than 16 hours</strong>" (1000x less) (<a href="https://arxiv.org/abs/1802.03268">paper</a>)</p></li></ul><p>We could argue why the same trend may not be true for this era of LLMs. After all, new techniques can also be scaled up and scale ultimately prevails as we know. In addition, the current trend of closed-source models makes it harder to build on them.</p><p>On the other hand, new powerful open-source models are still released regularly<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>. Companies are also incentivized to invest in the development of smaller models in order to reduce inference cost. Finally, we are starting to see the limits of scale on the horizon: recent LLMs are reaching the limits of text data online and repeating data eventually leads to diminishing returns (<a href="https://arxiv.org/abs/2305.16264">Muennighoff et al., 2023</a>) while Moore&#8217;s law is <a href="https://www.investopedia.com/terms/m/mooreslaw.asp">approaching its physical limits</a>.</p><p>There are already recent examples that require a fraction of compute by using new methods and insights, demonstrating that this trend also holds in the era of LLMs:</p><ul><li><p>FlashAttention (<a href="https://arxiv.org/abs/2205.14135">Dao et al., 2022</a>) provides drastic speedups over standard attention through clever hardware optimization.</p></li><li><p>Parameter-efficient fine-tuning methods (see our <a href="https://tinyurl.com/modular-fine-tuning-tutorial?ref=ruder.io">EMNLP 2022 tutorial</a> for an overview) including <a href="https://github.com/adapter-hub/adapters">adapters</a> such as LoRA (<a href="https://arxiv.org/abs/2106.09685">Hu et al., 2021</a>) and QLoRA (<a href="https://arxiv.org/abs/2305.14314">Dettmers et al., 2023</a>) enable fine-tuning LLMs on a single GPU.</p></li><li><p><a href="https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/">Phi-2</a>, a new 2.7B-parameter LLM released last week matches or outperforms models up to 25x its size.</p></li></ul><p>In the near term, the largest models using the most compute will continue to be the most capable. <strong>However, there remains a lot of room for innovation by focusing on strong smaller models and on areas where compute requirements will inexorably be eroded by research progress.</strong></p><p>While LLM projects typically require an exorbitant amount of resources, it is important to remind ourselves that <strong>research does not need to assemble full-fledged massively expensive systems in order to have impact</strong>. Chris Manning made the nice analogy in his EMNLP 2023 keynote that in the same vein, aerospace engineering students are not expected to engineer a new airplane during their studies.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!k77e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd6bd6fa4-bfbf-4af2-b66d-44417de1ce42_2560x1447.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!k77e!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd6bd6fa4-bfbf-4af2-b66d-44417de1ce42_2560x1447.webp 424w, https://substackcdn.com/image/fetch/$s_!k77e!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd6bd6fa4-bfbf-4af2-b66d-44417de1ce42_2560x1447.webp 848w, https://substackcdn.com/image/fetch/$s_!k77e!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd6bd6fa4-bfbf-4af2-b66d-44417de1ce42_2560x1447.webp 1272w, https://substackcdn.com/image/fetch/$s_!k77e!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd6bd6fa4-bfbf-4af2-b66d-44417de1ce42_2560x1447.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!k77e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd6bd6fa4-bfbf-4af2-b66d-44417de1ce42_2560x1447.webp" width="1456" height="823" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d6bd6fa4-bfbf-4af2-b66d-44417de1ce42_2560x1447.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:823,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:173950,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!k77e!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd6bd6fa4-bfbf-4af2-b66d-44417de1ce42_2560x1447.webp 424w, https://substackcdn.com/image/fetch/$s_!k77e!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd6bd6fa4-bfbf-4af2-b66d-44417de1ce42_2560x1447.webp 848w, https://substackcdn.com/image/fetch/$s_!k77e!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd6bd6fa4-bfbf-4af2-b66d-44417de1ce42_2560x1447.webp 1272w, https://substackcdn.com/image/fetch/$s_!k77e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd6bd6fa4-bfbf-4af2-b66d-44417de1ce42_2560x1447.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Members of the MIT aerospace engineering lab prepare a model of a super-efficient commercial aircraft for testing.</figcaption></figure></div><p>With that in mind, let&#8217;s look at five important research areas that require less compute.</p><div><hr></div><h2>1.  Efficient Methods</h2><p>Rather than waiting for compute costs to go down, making LLMs more efficient can have a wide impact. When we talk about efficiency, we often think about making the model architecture itself more efficient. In fact, most works on efficient Transformers focused on a specific component, the attention mechanism (<a href="https://arxiv.org/abs/2009.06732">Tay et al., 2022</a>).</p><p><strong>However, when thinking about efficiency, it is useful to consider the entire LLM stack.</strong> Important components ripe for improvement are:</p><ol><li><p><strong>Data collection and preprocessing</strong>: improving data efficiency by better filtering and data selection.</p></li><li><p><strong>Model input</strong>: faster, more informed tokenization; better word representations via character-level modeling</p></li><li><p><strong>Model architecture</strong>: better scaling towards long-range sequences; more effective use of memory</p></li><li><p><strong>Training</strong>: more efficient methods to train small-scale LLMs via more effective distillation, better learning rate schedules and restarts, (partial) model compression, model surgery, etc.</p></li><li><p><strong>Downstream task adaptation</strong>: improved parameter-efficient fine-tuning; automatic prompt and chain-of-thought design; <a href="https://www.modulardeeplearning.com/">modular methods</a>; improved RLHF</p></li><li><p><strong>Inference</strong>: early predictions; prompt compression; human-in-the-loop interactions</p></li><li><p><strong>Data annotation</strong>: model-in-the-loop annotation; automatic arbitration and consolidation of annotations</p></li><li><p><strong>Evaluation</strong>: efficient automatic metrics; efficient benchmarks</p></li></ol><p>Given the wide range of LLM applications, it is increasingly <strong>important to consider the &#8216;human&#8217; part in efficiency</strong>: from annotation, to learning from human preferences, to interacting with users, can we make the stages where human and LLM data intersect more efficient and reliable?</p><p><strong>Sparsity and low-rank approximations are two general principles</strong> that have been applied in a wide range of efficient methods (see our <a href="https://arxiv.org/abs/2302.11529">modular deep learning survey</a> for an overview) and are thus useful sources of inspiration: are there components that are modeled with an excess amount of parameters that can be approximated instead? Are there computations involving multiple steps that can be shortened?</p><p>In the age of LLMs, the clearest indicator that an efficient method works is that it reduces the coefficient (in other words, lowers the slope) of the corresponding scaling law, as seen for instance in <a href="https://arxiv.org/abs/2203.15556">Hoffmann et al. (2022)</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GVQS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GVQS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png 424w, https://substackcdn.com/image/fetch/$s_!GVQS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png 848w, https://substackcdn.com/image/fetch/$s_!GVQS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png 1272w, https://substackcdn.com/image/fetch/$s_!GVQS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GVQS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png" width="1456" height="562" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:562,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:173827,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GVQS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png 424w, https://substackcdn.com/image/fetch/$s_!GVQS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png 848w, https://substackcdn.com/image/fetch/$s_!GVQS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png 1272w, https://substackcdn.com/image/fetch/$s_!GVQS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe230fdb3-f786-4328-9065-b226e80bea6f_2085x805.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">More compute-efficient training leads to improved scaling of the Chinchilla model (<a href="https://arxiv.org/abs/2203.15556">Hoffmann et al., 2022</a>) compared to the predictions of the earlier scaling law by <a href="https://arxiv.org/abs/2001.08361">Kaplan et al. (2020)</a>. </figcaption></figure></div><p>But how can we validate a scaling law without massive compute? By prioritizing experimentation in small-scale regimes.</p><div><hr></div><h2>2.  Small-scale Problems</h2><p>While it is generally prohibitive to apply a new method directly to the largest model, using it on a smaller, representative model can serve as a useful prototype and proof of concept.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> These days in particular, <strong>one should not underestimate the pace of the ML and NLP community, which is receptive to and quick-to-adopt compelling new ideas.</strong></p><p>For instance, the recently proposed DPO method (<a href="https://arxiv.org/abs/2305.18290">Rafailov et al., 2023</a>) used a relatively small-scale experimental setting in the paper (GPT-2-large fine-tuned on IMDb reviews, among others). As the code was <a href="https://github.com/eric-mitchell/direct-preference-optimization">open-sourced</a> and compatible with common LLM frameworks, community members quickly applied it to more recent models such as <a href="https://huggingface.co/blog/dpo-trl">Llama-2</a> and <a href="https://huggingface.co/blog/alvarobartt/notus-7b-v1">Zephyr</a>.</p><p>Expect to see more of this mode of operation: <strong>academic researchers developing new methods that&#8212;after small-scale validation&#8212;are shared with the community for further experimentation and scaling up.</strong></p><p>Methods can also be developed on benchmarks that measure compute and sample efficiency and are designed with compute constraints in mind. Examples include the BabyLM Challenge (<a href="https://arxiv.org/abs/2301.11796">Warstadt et al., 2023</a>)&#8212;which focuses on sample-efficient pre-training on a developmentally plausible corpus of 10M and 100M tokens&#8212;and the Languini Kitchen (<a href="https://arxiv.org/abs/2309.11197">Stani&#263; et al., 2023</a>), which compares models based on equivalent compute.</p><p>Another setting where a focus on a small scale is increasingly valuable is analysis and model understanding. Through pre-training, models learn a wide array of natural language understanding capabilities&#8212;but under exactly what conditions these capabilities emerge remains unclear.</p><p>Large-scale pre-training, due to the massive nature of most of the components involved, mostly resists a controlled examination.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> <strong>Instead, controlled small and synthetic settings that allow probing of specific hypotheses will be increasingly important to understand how LLMs learn and acquire capabilities.</strong> Such settings can include synthetic language such as bigram data (<a href="https://arxiv.org/abs/2306.00802">Bietti et al., 2023</a>) or &#8220;fake&#8221; English (<a href="https://openreview.net/forum?id=HJeT3yrtDr">K et al., 2020</a>), highly curated and domain-specific data, and data satisfying certain (distributional) characteristics; as well as more interpretable models such as small Transformers, backpack language models (<a href="https://aclanthology.org/2023.acl-long.506/">Hewitt et al., 2023</a>), and neural additive models (<a href="https://arxiv.org/abs/2004.13912">Agarwal et al., 2021</a>).</p><p>LLM mechanisms whose emergence is still poorly understood include the following:</p><ul><li><p><strong>in-context learning</strong>: &#8216;burstiness&#8217; and the highly skewed distribution of language data are important (<a href="https://arxiv.org/abs/2205.05055">Chan et al., 2022</a>) but the in-context learning ability can also disappear again during training (<a href="https://arxiv.org/abs/2311.08360">Singh et al., 2023</a>)</p></li><li><p><strong>chain-of-thought prompting</strong>: local structure in the training data is important (<a href="https://openreview.net/pdf?id=rcXXNFVlEn">Prystawski et al., 2023</a>) but we don&#8217;t know how this relates to natural language data</p></li><li><p><strong>cross-lingual generalization</strong>: limited parameters, shared special tokens, shared position embeddings, and a common masking strategy contribute to multilinguality (<a href="https://aclanthology.org/2020.acl-main.421/">Artetxe et al., 2019</a>; <a href="https://aclanthology.org/2020.emnlp-main.358/">Dufter &amp; Sch&#252;tze, 2020</a>) but it is unclear how this extends to diverse natural language data and typologically diverse languages</p></li><li><p><strong>other types of emerging abilities</strong> (see for instance <a href="https://openreview.net/pdf?id=ITw9edRDlD">Schaeffer et al., 2023</a>)</p></li></ul><p>Rather than trying to make large-scale settings smaller to reduce the amount of compute necessary to study them, we can also focus on settings that are intrinsically small-scale due to constraints on the data available.</p><div><hr></div><h2>3.  Data-constrained Settings</h2><p>While the largest LLMs are pre-trained on trillions of tokens, the downstream applications we would like to apply them to are often more limited in the data available to them.</p><p>This is true for many interdisciplinary areas such as NLP for Science, Education, Law, and Medicine. In many of these domains, there is very little high-quality data easily accessible online. <strong>LLMs thus must be combined with domain-specific strategies to achieve the biggest impact.</strong> See <a href="https://arxiv.org/abs/2310.20633">Li et al. (2023)</a> for a brief review of directions in NLP+X applications.</p><p>Another area where data is notoriously limited is multilinguality. <strong>For many languages, the amount of text data online is limited&#8212;but data may be available in other formats such as lexicons, undigitized books, podcasts, and videos.</strong> This requires new strategies to collect&#8212;and create&#8212;high-quality data. Furthermore, many languages and dialects are more commonly spoken than written, which makes multi-modal models important to serve such languages.</p><p><strong>As we reach the limits of data available online, even &#8220;high-resource&#8221; languages will face data constraints.</strong> New research will need to engage with these constraints rather than assuming an infinite-scale setting.</p><p>While few-shot prompting enables seamless application to many downstream tasks, it is insufficient to teach a model about the nuances of more complex applications and is <a href="https://docs.google.com/presentation/d/1seHOJ7B0bQEPJ3LBW5VmruMCILiVRoPb8nmU2OS-Eqc/edit?ref=ruder.io#slide=id.g1a37bfe6b5e_3_92">limited in other ways</a>. Alternatively, parameter-efficient fine-tuning enables a more holistic adaptation using little compute. Such fine-tuning&#8212;when updates are constrained to a subset of model parameters&#8212;gives rise to <a href="https://arxiv.org/abs/2302.11529">modular models</a>.</p><p>Given the diversity of LLM application areas and capabilities to master, another interesting direction is thus to leverage multiple modular &#8216;experts&#8217; by learning to disentangle and combine the skills and knowledge learned across different domains.</p><p>Such modeling advances, however, are of little use if we do not have reliable means to evaluate them.</p><div><hr></div><h2>4.  Evaluation</h2><blockquote><p><em>"[...] benchmarks shape a field, for better or worse. Good benchmarks are in alignment with real applications, but bad benchmarks are not, forcing engineers to choose between making changes that help end users or making changes that only help with marketing."&#8212;David A. Patterson; foreword to <a href="https://www.springer.com/gp/book/9783030417048?ref=ruder.io">Systems Benchmarking (2020)</a></em></p></blockquote><p>In 2021, a common sentiment was that NLP models had outpaced the benchmarks to test for them. I reviewed the situation in <a href="https://www.ruder.io/nlp-benchmarking/">this article</a>; not much has changed since then. More recent benchmarks designed to evaluate LLMs such as <a href="https://crfm.stanford.edu/helm/latest/#/leaderboard">HELM</a> (<a href="https://arxiv.org/abs/2211.09110">Liang et al., 2022</a>) and Super-NaturalInstructions (<a href="https://aclanthology.org/2022.emnlp-main.340/">Wang et al., 2022</a>) still mainly consist of standard NLP tasks&#8212;most of them sentence-level&#8212;while others such as MMLU (<a href="https://arxiv.org/abs/2009.03300">Hendrycks et al., 2021</a>) and AGIEval (<a href="https://arxiv.org/abs/2304.06364">Zhong et al., 2023</a>) focus on exams. These benchmarks do not reflect the diverse range of tasks where we would like to apply LLMs.</p><p>Another phenomenon to be aware of is leaderboard contamination: benchmark data that is available online is likely to have been included in the pre-training data of LLMs, making evaluation unreliable. Benchmarks should thus keep evaluation data secret or receive regular updates.</p><blockquote><p><em>"When you can measure what you are speaking of and express it in numbers, you know that on which you are discussing. But when you cannot measure it and express it in numbers, your knowledge is of a very meagre and unsatisfactory kind."</em>&#8212;Lord Kelvin</p></blockquote><p>In addition, existing automatic metrics are ill-suited for more complex downstream applications and open-ended natural language generation tasks. LLMs can be incorporated into automatic metrics (<a href="https://aclanthology.org/2023.emnlp-main.153/">Liu et al., 2023</a>) but one must be aware of&#8212;and mitigate&#8212;their biases. For complex tasks, it may be useful to decompose them into subtasks that are easier to evaluate, for instance, via behavioral tests (<a href="https://aclanthology.org/2023.acl-long.396/">Hlavnova &amp; Ruder, 2023</a>).</p><p>As applications become more elaborate, even human evaluation, traditionally perceived to be the gold standard for any data, becomes less reliable. Disagreements may be less an indicator of &#8216;annotation noise&#8217; and rather a sign of different perspectives (<a href="https://aclanthology.org/Q19-1043/">Pavlick &amp; Kwiatkowski, 2019</a>). For specialized applications, only domain experts may be qualified enough to provide accurate feedback. Leveraging and aggregating the feedback of a diverse set of annotators from different backgrounds is thus more important than ever.</p><div><hr></div><h2>5.  Reasoning</h2><p>Reasoning requires the use of logic to draw conclusions from new and existing information to arrive at a conclusion. With LLMs demonstrating surprising arithmetic and logical reasoning abilities, reasoning has received renewed attention and was well-represented in NeurIPS 2023 papers (see my previous newsletter): </p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;4529fb01-2ab6-4fda-9aad-2d71e51ab2c7&quot;,&quot;caption&quot;:&quot;NeurIPS 2023, arguably this year&#8217;s biggest AI conference takes place in two weeks from Dec 10&#8211;16 in New Orleans. 3586 papers were accepted to the conference, which are available online. In this newsletter, I&#8217;ll discuss a selection of 20 papers related to natural language processing (NLP) that caught my eye, with a focus on oral and spotlight papers. Here&#8230;&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;&#128196; NeurIPS 2023 Primer&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:7965403,&quot;name&quot;:&quot;Sebastian Ruder&quot;,&quot;bio&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/17fdd4c3-a575-4fe4-b58e-d876b78bfe2f_2416x2416.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2023-12-01T15:51:54.818Z&quot;,&quot;cover_image&quot;:null,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://nlpnewsletter.substack.com/p/neurips-2023-primer&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:138865653,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:43,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;NLP News&quot;,&quot;publication_logo_url&quot;:&quot;https://s3.amazonaws.com/revue/profiles/images/000/027/670/thumb/sebastian_ruder_profile_photo_square.jpg?1616408613&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><p>Given that LLMs frequently hallucinate and struggle to generate code or plans that are directly executable, augmenting them with external tools or small domain-specific models is a promising direction to make them more robust. For instance, Parsel (<a href="https://nlpnewsletter.substack.com/p/neurips-2023-primer">Zelikman et al., 2023</a>) decomposes a code generation tasks into LLM-generated subfunctions that can be tested against input-output constraints using a code execution module.</p><p>Many complex real-world applications require different forms of reasoning so evaluating models&#8217; reasoning abilities in realistic scenarios is an important challenge. Given that many real-world problems require weighing different options and preferences, it will be crucial to enable LLMs to present different solutions to users and to incorporate different cultural backgrounds into their decision-making. <a href="https://arxiv.org/abs/2305.12544">Ignat et al. (2023)</a> highlight other interesting research directions related to reasoning.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">NLP News is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>This post presented a selection of five research directions that are particularly important in my opinion&#8212;but in truth there are a plethora of potential opportunities to explore (see the other reviews at the beginning of this post). Now is the time to look beyond standard NLP tasks and be ambitious. After all:</p><blockquote><p><em>&#8220;Shoot for the moon. Even if you miss, you'll land among the stars.&#8221;<br>&#8212;Norman Vincent Peale</em></p></blockquote><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Also known as <a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">Rich Sutton&#8217;s bitter lesson</a>, in other words, &#8220;<em>the great power of general purpose methods, of methods that continue to scale with increased computation even as the available computation becomes very great</em>&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>I shared &#8216;<a href="https://www.ruder.io/requests-for-research/">requests for research</a>&#8217; in 2018 where each task required a few GPUs at most for training a neural network from scratch.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>For an overview of NLP milestones until 2018, check out my <a href="https://www.ruder.io/a-review-of-the-recent-history-of-nlp/">&#8220;Review of the Neural History of NLP&#8221;</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Hugging Face referred to <a href="https://huggingface.co/blog/2023-in-llms">2023 as the year of open LLMs</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>For example, we recently trained mmT5 (<a href="https://arxiv.org/abs/2305.14224">Pfeiffer et al., 2023</a>), a modular version of multilingual T5 that dramatically outperforms its counterpart at small and base parameter sizes, demonstrating the benefits of modularity at this scale.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Nevertheless, certain areas such as memorization can be studied during large-scale pre-training by modifying the pre-training scheme appropriately, for instance via the insertion of &#8216;canaries&#8217; as in PaLM 2 (<a href="https://arxiv.org/abs/2305.10403">Anil et al., 2023</a>).</p></div></div>]]></content:encoded></item><item><title><![CDATA[📄 EMNLP 2023 Primer]]></title><description><![CDATA[In this newsletter, I&#8217;ll discuss a selection of exciting papers and workshops I&#8217;m looking forward to at EMNLP 2023 and the trends I observed.]]></description><link>https://newsletter.ruder.io/p/emnlp-2023-primer</link><guid isPermaLink="false">https://newsletter.ruder.io/p/emnlp-2023-primer</guid><pubDate>Tue, 05 Dec 2023 07:36:08 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b6eb16fc-914b-404e-b5a8-df37ae73f090_3584x2240.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><a href="https://2023.emnlp.org/">EMNLP 2023</a>, one of the biggest NLP conferences takes place this week from Dec 6&#8211;10 in Singapore.</p><p>In this newsletter, I&#8217;ll discuss a selection of exciting papers and workshops I&#8217;m looking forward to at the conference. Here are the main trends I observed (based on the small sample of papers I discuss here and those I came across online):</p><ol><li><p><strong>Instruction-tuned LMs and LLMs are everywhere.</strong> Similar to <a href="https://www.ruder.io/acl-2021-highlights/">earlier years</a> where BERT was ubiquitous, instruction-tuned language models (LMs) and large language models (LLMs) are used in almost every paper.</p></li><li><p><strong>Evaluation based on LLMs is increasingly common.</strong> While some papers employ automatic evaluation based on GPT-4, new metrics that are proposed are based on LLMs in zero-shot prompted or fine-tuned settings.</p></li><li><p><strong>Prompt usage is getting more creative.</strong> Beyond a standard prompt template, prompts are getting increasingly complex and specialized to the desired setting. Techniques such as <a href="https://arxiv.org/abs/2201.11903">chain-of-thought prompting</a> are common tools. </p></li><li><p><strong>Multilinguality&nbsp;is increasingly popular.</strong> I came across a substantial number of papers studying multilingual settings, which indicates that LLMs are still limited in non-English settings and that making LLMs more multilingual is an important direction. </p></li></ol><p>On the other hand, I did not come across many papers that tried to analyze LLM properties (using a synthetic setup, for instance) or that used external models or tools to augment LLMs (<em>please point me to papers that I missed</em>). This is in contrast to NeurIPS 2023 where such papers were more common (see the below newsletter).</p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;34bdeed0-343f-4f04-ad63-341ba765c114&quot;,&quot;caption&quot;:&quot;NeurIPS 2023, arguably this year&#8217;s biggest AI conference takes place in two weeks from Dec 10&#8211;16 in New Orleans. 3586 papers were accepted to the conference, which are available online. In this newsletter, I&#8217;ll discuss a selection of 20 papers related to natural language processing (NLP) that caught my eye, with a focus on oral and spotlight papers. Here&#8230;&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;&#128196; NeurIPS 2023 Primer&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:7965403,&quot;name&quot;:&quot;Sebastian Ruder&quot;,&quot;bio&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/17fdd4c3-a575-4fe4-b58e-d876b78bfe2f_2416x2416.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2023-12-01T15:51:54.818Z&quot;,&quot;cover_image&quot;:null,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://nlpnewsletter.substack.com/p/neurips-2023-primer&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:138865653,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:34,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;NLP News&quot;,&quot;publication_logo_url&quot;:&quot;https://s3.amazonaws.com/revue/profiles/images/000/027/670/thumb/sebastian_ruder_profile_photo_square.jpg?1616408613&quot;,&quot;belowTheFold&quot;:false,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><p>I&#8217;ll be attending the conference in-person so say &#8220;hi&#8221; if you&#8217;re there. </p><p><em>Papers with a &#8224; are presented in Findings of EMNLP (rather than the main conference). I am an author/co-organizer on papers/events indicated with a *.</em></p><h3>Workshops</h3><p><strong><a href="https://genbench.org/workshop/">GenBench Workshop</a>, Dec 6.</strong>  Generalization is crucial to ensure robust model behavior but how good generalization looks like and how it should be evaluated is still not well understood. The GenBench workshop on (benchmarking) generalization in NLP aims to catalyse research on generalization and how to measure it in the NLP community. Accepted papers study generalization or are <a href="https://github.com/google/BIG-bench">BIG-bench</a>-style <a href="https://genbench.org/cbt/">collaborative benchmarking tasks</a> (CBT). The <a href="https://genbench.org/workshop_programme/">program</a> consists of invited talks, CBT spotlights as well as oral presentations and posters.</p><p><strong><a href="https://nlposs.github.io/2023/index.html">Workshop for NLP Open-Source Software (NLP-OSS)</a>, Dec 6.</strong>  In light of the increasing number of closed-source LLMs, it is important to continue to promote an open culture of sharing knowledge, data, and software, from which the NLP community has benefited greatly. This workshop aims to further the sharing of insights regarding the creation and development of NLP open-source software. Invited talks feature important NLP open-source projects including <a href="https://github.com/CarperAI/trlx">trlX</a>, a framework for large-scale open-source RLHF and <a href="https://github.com/aisingapore/sealion">SEA-LION</a>, LLMs pre-trained for Southeast Asian languages.</p><p><strong><a href="https://www.bigpictureworkshop.com/">The Big Picture Workshop: Crafting a Research Narrative</a>, Dec 7</strong>*.&nbsp; In research, we &#8220;stand on the shoulders of giants&#8221;. However, given the number and rapid pace of published papers, it has become increasingly difficult, to recognize the larger story to which a paper is connected. The Big Picture Workshop aims to explore and distill such broader research narratives. We have a <a href="https://www.bigpictureworkshop.com/accepted-papers">diverse set of accepted papers</a> that provide insightful syntheses of different threads of research. On the workshop day, we&#8217;ll try out a new presentation format where we have researchers from different groups working on the same topic critically reflect on and discuss their work.</p><p><strong><a href="https://sigtyp.github.io/ws2023-mrl.html">Multilingual Representation Learning Workshop (MRL)</a>, Dec 7*</strong>.  This workshop provides a forum to discuss work to improve NLP in low-resource and under-represented languages. The large number of accepted papers and Findings papers explore a diverse set of methods, from meta-learning to tokenization and instruction tuning. In addition, shared task on multilingual multi-task information retrieval provided new data for NER and QA for a typologically diverse set of languages. The workshop day is jam-packed with excellent invited talks, poster, shared task and best paper sessions.</p><h3>Unanswerability and attribution in QA</h3><p>In question answering (QA), a crucial challenge for current LLMs is hallucinating answers. A scenario where such hallucinations are common is when questions do not have an answer. To deal with hallucinations, a promising strategy is to train the model to attribute the answer to relevant references.</p><p><strong><a href="https://arxiv.org/abs/2310.11877">The Curious Case of Hallucinatory (Un)answerability: Finding Truths in the Hidden States of Over-Confident Large Language Models (Slobodkin et al.)</a>.</strong>  This work shows that LLMs are aware of the concept of (un)answerability and that the representation of the first decoded token provides a strong indicator whether a question is answerable (removing this information in the first token significantly decreases performance). Furthermore, mentioning that the question is unanswerable in the prompt improves performance. <strong>These results highlight that developing better decoding methods will also help make LLMs more factual.</strong></p><p><strong><a href="https://arxiv.org/abs/2305.14332">Evaluating and Modeling Attribution for Cross-Lingual Question Answering (Muller et al.)</a>*.</strong>  This paper introduces attribution for cross-lingual question answering where the document supporting the generated answer may be in a different language than the question and answer. It creates the <a href="https://github.com/google-research/google-research/tree/master/xor_attriqa">XOR-AttriQA dataset</a> to measure attribution of SOTA QA models across 5 languages. Surprisingly, a large portion of generated answers are not attributable to any retrieved passage (up to 47% of <em>correctly</em> predicted answers in Japanese are not attributable). <strong>Current QA systems are thus often right but without any evidence, making them untrustworthy. </strong>Multilingual LLMs can be used to accurately detect attribution (which can complement string-based evaluation metrics) and can be used to rerank generated answers, improving QA performance. <strong>Key research directions are a) improving retrieval of cross-lingual passages and b) designing robust LLM-based metrics for QA evaluation.</strong></p><h3>Instruction tuning</h3><p>Instruction tuning is a common way to improve LLMs for downstream settings and to align them to human behavior. However, current instruction tuning datasets still have their limitations (see the below newsletters for an overview). </p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;3b14f3fa-4a60-4888-98bd-ae852d0f70be&quot;,&quot;caption&quot;:&quot;NLP and ML have gone through several phases of how models are trained in recent years. With the arrival of pre-trained models such as BERT, fine-tuning pre-trained models for downstream tasks became the norm. The increasing capabilities of ever larger models then enabled in-context learning via prompting. Recently, instruction tuning has become the newe&#8230;&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;&#129489;&#8205;&#127979; Instruction Tuning Vol. 1&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:7965403,&quot;name&quot;:&quot;Sebastian Ruder&quot;,&quot;bio&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/17fdd4c3-a575-4fe4-b58e-d876b78bfe2f_2416x2416.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2023-10-04T08:00:22.414Z&quot;,&quot;cover_image&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://nlpnewsletter.substack.com/p/instruction-tuning-vol-1&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:136684903,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:52,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;NLP News&quot;,&quot;publication_logo_url&quot;:&quot;https://s3.amazonaws.com/revue/profiles/images/000/027/670/thumb/sebastian_ruder_profile_photo_square.jpg?1616408613&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;b3ec4ddf-6c9b-4606-99e7-5c467fcccfc0&quot;,&quot;caption&quot;:&quot;Last month, we covered the first generation of instruction-tuning datasets that have been mainly based on existing NLP tasks. This month, we cover the latest datasets that are now much closer to real-world use cases&#8212;but still have their limitations!&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;Instruction Tuning Vol. 2&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:7965403,&quot;name&quot;:&quot;Sebastian Ruder&quot;,&quot;bio&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/17fdd4c3-a575-4fe4-b58e-d876b78bfe2f_2416x2416.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2023-11-15T10:13:42.126Z&quot;,&quot;cover_image&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://nlpnewsletter.substack.com/p/instruction-tuning-vol-2&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:137635802,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:19,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;NLP News&quot;,&quot;publication_logo_url&quot;:&quot;https://s3.amazonaws.com/revue/profiles/images/000/027/670/thumb/sebastian_ruder_profile_photo_square.jpg?1616408613&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><p><strong><a href="https://arxiv.org/abs/2305.14045">The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning (Kim et al.).</a></strong>  This paper introduces the Chain-of-Thought (CoT) Collection, which augments the <a href="https://arxiv.org/abs/2301.13688">Flan collection</a> covering 1,060 tasks with 1.84M chain-of-thought rationales. This amount of chain-of-thought instruction tuning data is particularly useful for smaller LMs and improves their performance on reasoning tasks including BIG-bench Hard and the multilingual MGSM benchmark.&nbsp; </p><h3>Task Adaptation</h3><p>While LLMs achieve very strong performance in a zero-shot setting, it is necessary to fine-tune them on task data to achieve the best performance. Keeping models updated as the distribution changes and encoding task knowledge efficiently across many settings are key challenges in this area.</p><p><strong><a href="https://arxiv.org/abs/2305.15076">Meta-Learning Online Adaptation of Language Models (Hu et al.)</a>.</strong>  Keeping LLMs up-to-date is an important challenge as it is prohibitive to re-train these models. This paper hypothesizes that when continual fine-tuning a model on a stream of documents, the learning signal of important documents may be drowned out. To ameliorate this, the authors propose to meta-train a small model to reweigh the LM loss for each token during online fine-tuning in order to maximize the QA model&#8217;s performance after a single weighted update. They find that this dynamic weighting significantly outperforms standard fine-tuning and weighting heuristics.</p><p><strong><a href="https://arxiv.org/abs/2311.11077">Adapters: A Unified Library for Parameter-Efficient and Modular Transfer Learning (Poth et al.)*.</a></strong>  Full fine-tuning of LLMs has become prohibitive and requires parameter-efficient methods instead. This demo paper presents <a href="https://github.com/adapter-hub/adapters">Adapters</a>, a library for <a href="https://www.modulardeeplearning.com/">parameter-efficient and modular learning</a> with LLMs and the successor to <code>adapter-transformers</code>. Adapters integrates 10 diverse modular methods such as <a href="https://arxiv.org/abs/2104.08691">prompt tuning</a>, <a href="https://arxiv.org/abs/2101.00190">prefix tuning</a>, <a href="https://arxiv.org/abs/2106.04647">Compacter</a>, <a href="https://arxiv.org/abs/2106.09685">LoRA</a>, and <a href="https://proceedings.neurips.cc/paper_files/paper/2022/file/0cde695b83bd186c1fd456302888454c-Paper-Conference.pdf">(IA)&#179;</a> into 20 state-of-the-art models for NLP, vision, and multi-modal applications. <strong>It supports a range of operations on these modules such as grouping, stacking, fusing, splitting, and parallelizing, among others, which enable a variety of modeling approaches and research directions.</strong></p><p><strong><a href="about:blank">Outlier Dimensions Encode Task-Specific Knowledge (Rudman et al.).</a></strong>  This paper shows that outlier dimensions (dimensions with a variance that is significantly higher than the average) in LLMs persist during fine-tuning. They also find that just using the embedding value of such a high-variance dimension with a linear threshold can achieve performance similar to using the full model for some tasks and models. <strong>We already know that LLMs capture task knowledge in a low-dimensional subspace (see <a href="https://aclanthology.org/2021.acl-long.568/">Aghajanyan et al., 2021</a>, for instance)&#8212;but the observation that the subspace can be 1D for some settings can motivate the development of new efficient methods.</strong></p><h3>NLG Evaluation</h3><p>As LLMs are increasingly applied to generate natural language text, we need better metrics to evaluate their performance. One of the most promising directions is to use LLMs themselves as part of the metric, whether in a zero-shot setting or fine-tuned on relevant data.</p><p><strong><a href="https://arxiv.org/abs/2303.16634">G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment (Liu et al.).</a></strong>  This paper proposes G-Eval, a framework for NLG evaluation using LLMs as reference-free metrics.&nbsp;Given a description of the task and the evaluation criteria, they first generate a more detailed CoT-style description of the evaluation steps using an LLM. All descriptions are then concatenated with the input example and fed to the LLM. Rather than directly predicting a score for each evaluation criterion, the authors observed that they obtain better measurements if they instead take the sum of all candidate scores weighted with their probability. On summarization, the framework achieves a higher correlation with human judgements than existing metrics.</p><p><strong><a href="https://arxiv.org/abs/2305.14282">INSTRUCTSCORE: Explainable Text Generation Evaluation with Finegrained Feedback (Xu et al.).</a></strong>  This paper proposes InstructScore, a fine-grained reference-based explainable metric for NLG evaluation using LLMs. To fine-tune the LLM as the metric, they first collect unlabeled sentences. They then specify the number of errors, error types, and their severity labels for each sentence and ask GPT-4 to generate an incorrect sentence containing the errors matching the criteria and an explanation for each error. LLaMa is then fine-tuned on the generated data to identify and explain the errors in the incorrect sentence compared to the reference. LLaMa is further refined using feedback from GPT-4 regarding the correctness of the generated explanations. In practice, InstructScore achieved similar or higher correlation with human judgements than existing metrics on translation and NLG tasks.</p><h3>Multilingual Models</h3><p>While current LLMs excel on many tasks for English, performance is still much worse on languages with limited data. We thus require models that perform well for such languages and new methods to effectively scale models to these languages.</p><p><strong><a href="https://arxiv.org/abs/2311.05640">FinGPT: Large Generative Models for a Small Language (Luukkonen et al.).</a></strong>  This paper is a comprehensive study of training LLMs for a small language (Finnish) including the collection of a diverse dataset, monolingual training at different model sizes (up to 13B parameters), adaptation of an existing multilingual language model (BLOOM) to the new language, and creation of a language-specific benchmark. The trained models outperform all previous models for Finnish while the language-adapted multilingual model outperforms the monolingual models. <strong>Overall, this is a nice blueprint of how LLMs can be trained for medium-resource languages.</strong></p><p><strong><a href="https://arxiv.org/abs/2305.14224">mmT5: Modular Multilingual Pre-Training Solves Source Language Hallucinations (Pfeiffer et al.)</a><a href="https://arxiv.org/abs/2305.11938">*&#8224;</a><a href="https://arxiv.org/abs/2305.14224">.</a></strong>  This paper proposes mmT5, the first modular multilingual generative model. The mT5-style model is pre-trained with language-specific modules and dramatically outperforms mT5 and at similar parameter sizes while matching or outperforming XLM-R. <strong>Importantly, the model&#8217;s modularity enables more direct control over its outputs.</strong> While mT5 generates text in the correct language in only 7% of cases for zero-shot cross-lingual summarization, mmT5 generates text in the correct language in 99% (!) of cases.</p><p><strong><a href="https://arxiv.org/abs/2301.10472">XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models (Liang et al.).</a></strong>  This paper proposes XLM-V, an XLM-R-style model covering 100 languages that is pre-trained with a 1M vocabulary. To create the vocabulary, vocabularies of languages are first clustered (<a href="https://aclanthology.org/2020.emnlp-main.367/">Chung et al., 2020</a>), clusters are allocated capacity corresponding to their average log probability (<a href="https://aclanthology.org/2021.emnlp-main.257/">Zheng et al., 2021</a>), and sentencepiece models are trained for each cluster and then combined. While pre-training with a 1M vocabulary is 2.5x slower than with a 250k vocabulary, the resulting model outperforms a (reimplemented) XLM-R. </p><p><strong><a href="https://arxiv.org/abs/2304.08865">Romanization-based Large-scale Adaptation of Multilingual Language Models (Purkayastha et al.)</a>*&#8224;</strong><em><strong>.</strong></em>  This paper explores the potential of large-scale transliteration to enable multilingual LMs to deal with under-represented languages. In particular, the paper romanizes (i.e., maps UTF-8 characters to Latin characters) text using <a href="http://( UTF-8 characters to Latin characters) text">uroman</a> across 14 diverse languages, which is then used to adapt multilingual LMs. Romanization is particularly useful in the most challenging setups: on languages with unseen scripts and with limited training data.</p><h3>Multilingual Datasets and Evaluation</h3><p>A key challenge for multilingual NLP is the lack of evaluation datasets and studies that accurately assess the performance of multilingual models. The creation of new datasets and the development of new evaluation measures and analyses is thus an important research direction.</p><p><strong><a href="https://arxiv.org/abs/2305.14235">Multilingual Large Language Models Are Not (Yet) Code-Switchers (Zhang et al.).</a></strong>  This paper evaluates LLMs on three code-switching tasks: sentiment analysis (English-{Spanish, Malayalam, Tamil}), translation (English-Hindi), and word-level language identification (English-Hindi, Standard Arabic-Egyptian Arabic). They observe that smaller fine-tuned multilingual LMs (XLM-R, mDeBERTa) still outperform zero-shot prompted LLMs on these tasks.</p><p><strong><a href="https://arxiv.org/abs/2305.11938">XTREME-UP: A User-Centric Scarce-Data Benchmark for Under-Represented Languages (Ruder et al.)*&#8224;</a>.</strong>  XTREME-UP is a new benchmark focusing on user-centric tasks in under-represented languages with realistic amounts of available data. The benchmark includes impactful multi-modal tasks such as ASR and OCR, which we make accessible for text-only models by providing baseline system outputs (in addition to the original audio and image inputs). We created new data for a range of different tasks and updated standard tasks such as QA and NER to make them more practically relevant. We find that multilingual fine-tuned models still outperform few-shot prompted models on most tasks and that character-level modeling is beneficial. Overall, there is still a lot of headroom left to improve performance on under-represented languages.</p><p><strong><a href="https://arxiv.org/abs/2310.14557">The Skipped Beat: A Study of Sociopragmatic Understanding in LLMs for 64 Languages (Zhang et al.).</a></strong>  This paper introduces SPARROW, a multilingual multi-task benchmark spanning 169 datasets from different online platforms to measure <em>sociopragmatic</em> understanding in LLMs (i.e., how well they perform on tasks related to social interactions such as sentiment analysis, emotion detection, etc). They observe that the fine-tuned models outperform zero-shot prompted models as well as ChatGPT. LLMs perform particularly poorly on humor and antisocial language detection and ChatGPT performs poorly across most languages in comparison to the best model.</p><p><strong><a href="https://arxiv.org/abs/2302.08956">AfriSenti: A Twitter Sentiment Analysis Benchmark for African Languages (Muhammad et al.)*.</a></strong>  This paper introduces AfriSenti, a sentiment analysis benchmark consisting of 110k+ tweets in 14 African languages. The dataset was used in the <a href="https://afrisenti-semeval.github.io/">AfriSenti SemEval-2023 Shared Task</a>. Data collection and annotation challenges included a lack of support for African languages by the Twitter API, lack of tone markings, frequent code-mixing and dialects, sarcasm and ambiguities, and a lack of annotators and a reliable Internet connection. The strongest model, <a href="https://aclanthology.org/2022.coling-1.382/">AfroXLM-R</a>, achieves 67.2 accuracy across all languages, leaving ample room for improvement.</p><p><strong><a href="https://arxiv.org/abs/2310.04928">Large Language Models Only Pass Primary School Exams in Indonesia: A Comprehensive Test on IndoMMLU (Koto et al.).</a></strong>  This paper introduces IndoMMLU, the first benchmark on Indonesian language and culture consisting of 15k questions from primary school to university entrance exams. Among the 24 evaluated models, GPT-3.5 is the only one that passes primary school exams while no LLM demonstrates familiarity with local Indonesian languages and culture. The language exams also enable assessing the level of Indonesian language proficiency. For grades 7 and above, GPT-3.5 fails to pass the exam while other models only pass grades 1&#8211;3.</p><p><strong><a href="https://arxiv.org/abs/2211.00142">TaTA: A Multilingual Table-to-Text Dataset for African Languages (Gehrmann et al.)*&#8224;.</a></strong>  This paper proposes <a href="https://github.com/google-research/url-nlp/tree/main/tata">Table-to-Text in African languages (TaTA)</a>, the first large multilingual table-to-text dataset with a focus on African languages. TaTA was created by transcribing figures and associated text in bilingual reports by the <a href="https://dhsprogram.com/">DHS Program</a>, which were then professionally translated to make the dataset fully parallel. We find that less than half of the outputs from an mT5-XXL-based model are understandable and attributable to the source data. We also observe that existing metrics perform poorly for multilingual table-to-text generation and introduce a new learned metric that achieves a high correlation with human judgements.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">NLP News is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>What papers did you find exciting at EMNLP 2023? Let me know in the comments.</em></p>]]></content:encoded></item><item><title><![CDATA[📄 NeurIPS 2023 Primer]]></title><description><![CDATA[A Round-up of 20 Exciting LLM-related Papers]]></description><link>https://newsletter.ruder.io/p/neurips-2023-primer</link><guid isPermaLink="false">https://newsletter.ruder.io/p/neurips-2023-primer</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Fri, 01 Dec 2023 15:51:54 GMT</pubDate><enclosure url="https://s3.amazonaws.com/revue/profiles/images/000/027/670/thumb/sebastian_ruder_profile_photo_square.jpg?1616408613" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><a href="https://nips.cc/">NeurIPS 2023</a>, arguably this year&#8217;s biggest AI conference takes place in two weeks from Dec 10&#8211;16 in New Orleans. <a href="https://blog.neurips.cc/2023/11/20/neurips-newsletter-october-2023/">3586 papers</a> were accepted to the conference, which are <a href="https://openreview.net/group?id=NeurIPS.cc/2023/Conference">available online</a>.</p><p>In this newsletter, I&#8217;ll discuss a selection of 20 papers related to natural language processing (NLP) that caught my eye, with a focus on oral and spotlight papers. Here are the main trends I observed:</p><ol><li><p><strong>Most NLP work at NeurIPS is related to large language models (LLMs).</strong> While there are some papers that do not employ LLMs or use a different setting (see <a href="https://openreview.net/forum?id=ez6Cb0ZGzG">Suhr &amp; Artzi</a> below, for instance), papers still presented their contributions in the context of LLMs.</p></li><li><p><strong>Synthetic setups to analyze LLM properties are becoming more common.</strong> This is because it is computationally prohibitive to run many different pre-training experiments. Investigated properties range from the emergence of in-context learning and learning using global statistics to chain-of-thought reasoning.</p></li><li><p><strong>Aligning models based on human preferences received a lot of attention.</strong> Papers particularly focused on improving RLHF and studying alignment to specific personality traits and beliefs.</p></li><li><p><strong>A comprehensive understanding of in-context learning still remains elusive. </strong>Papers studied different aspects of in-context learning such as whether it persists during training and using a Bayesian perspective. </p></li><li><p><strong>Reasoning is still challenging with current models.</strong> Papers focused on improving performance on various types of reasoning tasks including pragmatic, graph-based, algorithmic, compositional, and planning-based reasoning.</p></li><li><p><strong>External tools are increasingly used to improve LLMs&#8217; reasoning abilities.</strong> These range from external verifiers to code execution modules. </p></li></ol><p><em>Note that some of the methods proposed in these papers such as <a href="https://github.com/eric-mitchell/direct-preference-optimization">DPO</a> and <a href="https://github.com/artidoro/qlora">QLoRA</a> have already been successfully used in LLM applications.</em></p><div><hr></div><h3>Rethinking LLMs</h3><p>This is one of my favorite topics as these papers encourage us to rethink our fundamental assumptions regarding LLMs and provide new insights and perspectives on their inner workings.</p><p><strong><a href="https://openreview.net/forum?id=NiQTy0NW1L">Lexinvariant Language Models (Huang et al.).</a>  </strong>One of the fundamental characteristics of LMs that hasn&#8217;t changed since the <a href="https://www.jmlr.org/papers/volume3/bengio03a/bengio03a.pdf">first neural LM paper</a> is the one-to-one mapping between tokens and embeddings. This paper studies whether language modeling can also be done with models that are &#8216;lexinvariant&#8217;, i.e., that do not have fixed token embeddings but assign the <em>lexical permutation</em> of each sequence the same probability.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> This seems like a strong limitation&#8212;but it can serve as a useful inductive bias for recovering substitution cyphers (via an MLP probe) and in-context symbol manipulation. In practice, tokens are encoded using random Gaussian vectors and sampled so that the same token has the same representation <em>within</em> a sequence but different representations <em>across</em> sequences. <strong>While this method is mainly of theoretical interest, using it as regularization by using random embeddings only for a subset of tokens improves results on some BIG-bench tasks.</strong></p><div><hr></div><h3>Learning from Human Feedback</h3><p>Given the proliferation of different pre-trained LLMs, researchers and practitioners are increasingly looking to improve the next step in the LLM pipeline: learning from human feedback, which is important for maximizing performance on downstream tasks but also for LLM alignment.</p><p><strong><a href="https://openreview.net/pdf?id=HPuSIXJaa9">Direct Preference Optimization: Your Language Model is Secretly a Reward Model (Rafailov et al.).</a>  </strong>Reinforcement learning from human feedback (RLHF) is the preferred approach to update LLMs to align with target preferences but is quite complex (it requires first training a reward model and then updating the LLM with RL based on the reward model) and can be unstable. This paper proposes Direct Preference Optimization (DPO), which shows that the same objective can be optimized via a simple classification-based objective on the preference data&#8212;without any RL! An important component of the objective is a dynamic, per-example importance weight. <strong>DPO has the potential to make aligning LLMs with human preferences much more seamless&#8212;and will thus be important for safety research.</strong></p><p><strong><a href="https://openreview.net/forum?id=CSbGXyCswu">Fine-Grained Human Feedback Gives Better Rewards for Language Model Training (Wu et al.).</a>  </strong>This paper addresses another limitation of RLHF, which does not allow the integration of more fine-grained feedback regarding which parts of the generated response are erroneous. The paper proposes Fine-grained RLHF, which <strong>a)</strong> uses a dense reward model (a reward for every output sentence rather than for the entire output) and <strong>b)</strong> incorporates multiple reward models for diverse feedback. They experiment on detoxification and long-form question answering where they see improved results compared to RLHF and supervised fine-tuning. Importantly, as providing human preference judgements for RLHF is a complex annotation task, providing more fine-grained feedback is actually <em>not</em> more time-intensive. <strong>Expect to see more approaches experimenting with various reward models at different granularities.</strong></p><p><strong><a href="https://openreview.net/forum?id=ez6Cb0ZGzG">Continual Learning for Instruction Following from Realtime Feedback (Suhr &amp; Artzi).</a></strong>  This paper tackles continual learning from human feedback in a collaborative 3D world environment. They demonstrate a simple approach using a contextual bandit to update the model&#8217;s policy using binary rewards. Over 11 rounds of training and deployment, instruction execution accuracy improves from 66.7% to 82.1%. Empirically, in this setting, the feedback data provides a similar amount of learning signal as the supervised data. <strong>While their setting differs from the standard text-based scenario, it provides a sketch for how instruction-following agents can continually learn from human feedback.</strong> In the future, we will likely see approaches that utilize more expressive feedback such as via natural language.</p><div><hr></div><h3>LLM Alignment</h3><p>In order to ensure that LLMs are most useful, it is crucial to align them with the specific guidelines, safety policies, personality traits and beliefs that are relevant for a given downstream setting. To do this, we first need to understand what tendencies LLMs already encode&#8212;and then develop methods to steer them appropriately.</p><p><strong><a href="https://openreview.net/forum?id=I9xE1Jsjfx">Evaluating and Inducing Personality in Pre-trained Language Models (Jiang et al.).</a>  </strong>This paper proposes to assess the personality of LLMs based on the <a href="https://en.wikipedia.org/wiki/Big_Five_personality_traits">Big Five personality traits</a> known from psychology. Building on existing questionnaires, they create multiple-choice question answering examples where the LLM must choose how accurately statements such as &#8220;You love to help others&#8221; describe it. Each statement is associated with a personality trait. Crucially, it is less important whether a model scores highly on a specific trait but whether it exhibits a <em>consistent</em> personality, that is, whether it responds similarly to all questions associated with the trait. Only the largest models exhibit consistent personality traits that are similar to those of humans. <strong>It will be interesting to better understand under what conditions personality traits emerge and how consistent personalities can be encoded in smaller models.</strong></p><p><strong><a href="https://openreview.net/forum?id=CbsJ53LdKc">In-Context Impersonation Reveals Large Language Models' Strengths and Biases (Salewski et al.).</a></strong>  There has been a lot of anecdotal evidence that prompting LLMs to impersonate domain experts (e.g., &#8220;you are an expert programmer&#8221;, etc) improved models&#8217; capabilities. This paper studies such in-context impersonation across different tasks and finds indeed that LLMs impersonating domain experts perform better than LLMs impersonating non-domain experts. Impersonation is also useful to detect implicit biases. For instance, LLMs impersonating a man describe cars better than ones prompted to be a woman (based on CLIP&#8217;s ability to match an image to a category using the generated description of the category). <strong>Overall, impersonation is a useful tool to analyze LLMs&#8212;but may reinforce biases when used for (system) prompts.</strong></p><p><strong><a href="https://openreview.net/forum?id=O06z2G18me">Evaluating the Moral Beliefs Encoded in LLMs (Scherer et al.).</a>  </strong>This paper studies how moral beliefs are encoded in LLMs with regard to both high-ambiguity (&#8220;Should I tell a white lie?&#8221;) and low-ambiguity scenarios (&#8220;Should I stop for a pedestrian on the road?&#8221;) scenarios. They evaluate 28 (!) different LLMs and find that <strong>a)</strong> in unambiguous scenarios most models align with commonsense while in ambiguous cases, most models express uncertainty; <strong>b)</strong> models are sensitive to the wording of the question; and <strong>c)</strong> some models exhibit clear preferences in ambiguous scenarios&#8212;and closed-source models have similar preferences. <strong>While the evaluation relies on a heuristic mapping of output sequences to actions, <a href="https://github.com/ninodimontalcino/moralchoice">the data</a> is useful for further research of LLMs&#8217; moral beliefs.</strong></p><div><hr></div><h3>LLM Pre-training</h3><p>Pre-training is the most compute-intensive part of LLM pipelines and is thus harder to study at scale. Nevertheless, innovations such as new scaling laws improve our understanding of pre-training and inform future training runs.</p><p><strong><a href="https://openreview.net/pdf?id=j5BuTrEj35">Scaling Data-Constrained Language Models (Muennighoff et al.).</a>  </strong>As the amount of text online is limited, this paper investigates scaling laws in data-constrained regimes, in contrast to the scaling laws by <a href="https://arxiv.org/abs/2203.15556">Hoffmann et al. (2022)</a>, which focused on scaling without repeating data. The authors observe that training for up to 4 epochs on repeated data performs similarly to training on unique data. With more repetition, however, the value of additional training rapidly diminishes. In addition, augmenting the pre-training data with code meaningfully increases the pre-training data size. <strong>In sum, whenever we don&#8217;t have infinite amounts of pre-training data, we should train smaller models for more (up to 4) epochs.</strong></p><div><hr></div><h3>LLM Fine-tuning</h3><p>Fine-tuning large models with back-propagation is expensive so these papers propose methods to make fine-tuning more efficient, either using <a href="https://www.modulardeeplearning.com/">parameter-efficient fine-tuning</a> methods or without computing gradients (zeroth-order optimization). </p><p><strong><a href="https://openreview.net/pdf?id=OUIFPHEgJU">QLoRA: Efficient Finetuning of Quantized LLMs (Dettmers et al.).</a></strong>  This paper proposes QLoRA, a more memory-efficient (but <a href="https://lightning.ai/pages/community/lora-insights/">slower</a>) version of LoRA that uses several optimization tricks to save memory. They train a new model, Guanaco, that is fine-tuned only on a single GPU for 24h and outperforms previous models on the Vicuna benchmark. <strong>Overall, QLoRA enables using much fewer GPU memory for fine-tuning LLMs. </strong>Concurrently, other methods such as <a href="https://github.com/johnsmith0031/alpaca_lora_4bit">4-bit LoRA quantization</a> have been developed that achieve similar results.</p><p><strong><a href="https://openreview.net/pdf?id=Vota6rFhBQ">Fine-Tuning Language Models with Just Forward Passes (Malladi et al.).</a>  </strong>This paper proposes a memory-efficient zeroth-order optimizer (MeZO) as a more memory-efficient version of a classic zeroth-order optimizer that uses differences of loss values to estimate gradients. In practice, the method achieves similar performance to fine-tuning on several tasks but requires 100x more optimization steps (while being faster at each iteration). <strong>Nevertheless, it is surprising that such zeroth-order optimization works with very large models in the first place, demonstrating the robustness of such models, and is an interesting direction for future work.</strong></p><div><hr></div><h3>Emergent Abilities and In-context Learning</h3><p>Certain abilities of LLMs such as in-context learning and arithmetic reasoning have been shown to be present only in the largest models. It is still unclear how these abilities are acquired during training and what specific properties lead to their emergence, motivating many studies in this area.</p><p><strong><a href="https://openreview.net/pdf?id=ITw9edRDlD">Are Emergent Abilities of Large Language Models a Mirage? (Schaeffer et al.)</a>.  </strong>Emergent abilities are abilities that are present in large-scale models but not in smaller models and are hard to predict. Rather than being a product of models&#8217; scaling behavior, this paper argues that emergent abilities are mainly an artifact of the choice of metric used to evaluate them. Specifically, nonlinear and discontinuous metrics can lead to sharp and unpredictable changes in model performance. Indeed, the authors find that when accuracy is changed to a continuous metric for arithmetic tasks where emergent behavior was previously observed, performance improves smoothly instead. <strong>So while emergent abilities may still exist, they should be properly controlled and researchers should consider how the chosen metric interacts with the model.</strong></p><p><strong><a href="https://arxiv.org/abs/2311.08360">The Transient Nature of Emergent In-Context Learning in Transformers (Singh et al.).</a>   </strong><a href="https://arxiv.org/abs/2205.05055">Chan et al. (2022)</a> have previously shown that the distributional properties of language data (specifically, &#8216;burstiness&#8217; and a highly skewed distribution) play an important role in the emergence of in-context learning (ICL) in LLMs. Prior work also generally assumes that once the ICL ability has been acquired, it is retained by the model as learning progresses. This paper uses <a href="https://github.com/brendenlake/omniglot">Omniglot</a>, a synthetic image few-shot dataset to show cases where ICL emerges&#8212;only to be subsequently lost while the loss continues to decrease. On the other hand, L2 regularization seems to help the model retain its ICL ability. <strong>It is still unclear, however, how ICL emerges and if transience can be observed during LLM pre-training on real-world natural language data.</strong></p><p><strong><a href="https://openreview.net/pdf?id=rcXXNFVlEn">Why think step by step? Reasoning emerges from the locality of experience (Prystawski et al.).</a>  </strong>This paper investigates why and how chain-of-thought reasoning<a href="https://arxiv.org/abs/2201.11903"> (Wei et al., 2022)</a> is useful in LLMs using a <a href="https://github.com/benpry/why-think-step-by-step">synthetic setup</a>. Similar to <a href="https://arxiv.org/abs/2205.05055">Chan et al. (2022)</a>, they study distributional properties of the pretraining data. They find that chain-of-thought reasoning is only useful when the training data is locally structured. In other words, when examples are about closely connected topics as is common in natural language. They find that chain-of-thought reasoning is helpful because it incrementally chains local statistical dependencies that are frequently observed in training. <strong>It is still unclear, however, when chain-of-thought reasoning emerges during training and what are the properties of downstream tasks where it is most useful.</strong></p><p><strong><a href="https://arxiv.org/abs/2301.11916">Large Language Models Are Latent Variable Models: Explaining and Finding Good Demonstrations for In-Context Learning (Wang et al.).</a>  </strong>This paper frames in-context learning with LLMs as topic modeling where the generated tokens are conditioned on a latent topic (concept) variable, which captures format and task information. To make this computationally efficient, they use a smaller LM to learn the latent concepts via prompt tuning on the full demonstration data. They then select the examples that achieve the highest probability under the prompt-tuned model as demonstrations for in-context learning, which improves over other selection baselines. <strong>This is a further data point that shows that examples that are probable based on a latent concept of a task are useful demonstrations.</strong> This likely isn&#8217;t the full picture, however, and it will be interesting to see how this formulation relates to other data similarity and diversity measures.</p><p><strong><a href="https://openreview.net/forum?id=3X2EbBLNsk">Birth of a Transformer: A Memory Viewpoint (Bietti et al.).</a> </strong> This study investigates how LLMs learn in-context learning as well as the ability to use more general knowledge in a synthetic setup. The setup consists of sequences generated by a bigram LM where some bigrams require the local context to infer them while others require global statistics. They find that two-layer transformers (but not one-layer transformers) develop an induction head consisting of a &#8220;circuit&#8221; of two attention heads to predict in-context bigrams. They then freeze some of the layers to study the model&#8217;s training dynamics, finding that global bigrams are learned first and that the induction head learns appropriate memories in a top-down fashion. <strong>Overall, this paper sheds further light on how in-context learning can emerge in LMs.</strong></p><div><hr></div><h3>Reasoning</h3><p>Reasoning tasks that require the systematic chaining or composition of different pieces of information are one of the most important problems for current LLMs to solve. Their challenging nature and the diversity of domains where they are relevant makes them a fruitful area for research. </p><p><strong><a href="https://openreview.net/forum?id=X6dEqXIsEW">On the Planning Abilities of Large Language Models - A Critical Investigation (Valmeekam et al.).</a>  </strong>Planning and sequential decision making are important for a wide range of applications. This paper studies whether LLMs can generate simple plans for <a href="https://github.com/karthikv792/LLMs-Planning">commonsense planning tasks</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>. Used on their own, only 12% of the plans generated by the best LLM are directly executable. However, when combined with an automated planning algorithm that can identify and remove errors in the LLM-generated plan, LLMs do much better than using a random or empty initial plan. LLMs plans can also be improved via prompting based on feedback from an external verifier. On the other hand, the benefits of LLMs disappear when action names are unintelligible and cannot be easily inferred with common sense, which indicates a lack of abstract reasoning ability. <strong>Overall, while initial plans of LLMs can be useful as a starting point, LLM-based planning currently works best mainly in conjunction with external tools.</strong></p><p><strong><a href="https://openreview.net/forum?id=UDqHhbqYJV">Can Language Models Solve Graph Problems in Natural Language? (Wang et al.)</a>  </strong>The authors create <a href="https://github.com/Arthur-Heng/NLGraph">NLGraph</a>, a benchmark of graph-based reasoning problems described in natural language and evaluate LLMs on it. They find that LLMs demonstrate impressive preliminary graph reasoning abilities&#8212;37&#8211;58% above random baselines. However, in-context learning and advanced prompting strategies (chain-of-thought prompting and others) are mostly ineffective on more complex graph problems and LLMs are susceptible to spurious correlations. More specialized graph prompting strategies, however, improve results. <strong>Expect to see combinations of standard graph-based methods (such as those applied to <a href="https://paperswithcode.com/dataset/fb15k">FB15k</a>) and LLMs and research on methods scaling LLMs to larger graphs.</strong></p><p><strong><a href="https://openreview.net/forum?id=5bWW9Eop7l">The Goldilocks of Pragmatic Understanding: Fine-Tuning Strategy Matters for Implicature Resolution by LLMs (Ruis et al.).</a>  </strong>This paper studies whether LLMs exhibit a particular type of pragmatic inference, <a href="https://en.wikipedia.org/wiki/Implicature#:~:text=Article%20Talk,it%20is%20not%20literally%20expressed.">implicature</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. The authors evaluate whether models can understand implicature by measuring whether they assign a higher probability to a statement that contains the correct inference compared to the incorrect one. They find that both instruction tuning and chain-of-thought prompting are important for such pragmatic understanding and that the largest model, GPT-4 reaches human-level performance. <strong>We will likely see more work on different types of pragmatic understanding as these are crucial for seamless and human-like conversations.</strong></p><p><strong><a href="https://openreview.net/forum?id=qd9qcbVAwQ">Parsel&#128013;: Algorithmic Reasoning with Language Models by Composing Decompositions (Zelikman et al.).</a></strong>  This paper introduces <a href="https://github.com/ezelikman/parsel">Parsel</a>, a framework to implement complex programs with code LLMs. An LLM is first used to generate natural language function descriptions in a simple intermediate language. For each function description, the model then generates implementation candidates. The function candidates are then tested against input-output constraints and composed to form the final program. Using GPT-4 as LLM, this framework increased performance on HumanEval from 67% to 85%. <strong>This is a great example of how LLMs can be used as a building block together with domain-specific knowledge and tools for much improved performance. </strong>In addition, expect to see more work on breaking down complex decision problems into subproblems that are more easily solvable using LLMs.</p><p><strong><a href="https://openreview.net/forum?id=Fkckkr3ya8">Faith and Fate: Limits of Transformers on Compositionality (Dziri et al.).</a></strong> This paper investigates the compositional reasoning abilities of LLMs. They formulate compositional reasoning tasks as computation graphs in order to quantify their complexity. They find that full computation subgraphs appear significantly more frequently in the training data for correctly predicted test examples than for incorrectly predicted ones, indicating that models learn to match subgraphs rather than developing systematic reasoning skills. Using a scratchpad and grokking (training beyond overfitting) similarly do not improve performance on more complex problems. <strong>Overall, current LLMs still struggle with composing operations into correct reasoning paths.</strong></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">NLP News is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>Which NeurIPS papers did you find most exciting? Let me know in the comments.</em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Under such a model, the phrase <code>&#8220;a big banana&#8221;</code> has the same probability as the phrase <code>&#8220;e cop cekeke&#8221;</code> as both are the same given the permutation <code>{a: e, b: c, i: o, g: p, n:k}</code>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>These tasks consists of a problem domain, an initial state, and a goal state. The problem domain consists of a set of actions, for instance, to pick up blocks.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>An example of an implicature is the exchange &#8220;Did you go to the shop today?&#8221; &#8220;It was closed&#8221; where the second utterance implies &#8220;no&#8221; as the answer to the question.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Instruction Tuning Vol. 2]]></title><description><![CDATA[We cover the latest generation of instruction-tuning datasets that are much closer to real-world use cases&#8212;but still have their limitations!]]></description><link>https://newsletter.ruder.io/p/instruction-tuning-vol-2</link><guid isPermaLink="false">https://newsletter.ruder.io/p/instruction-tuning-vol-2</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Wed, 15 Nov 2023 10:13:42 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last month, we covered the first generation of instruction-tuning datasets that have been mainly based on existing NLP tasks.</p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;8d13b0ae-fa0b-4167-8032-58cc4b456880&quot;,&quot;caption&quot;:&quot;NLP and ML have gone through several phases of how models are trained in recent years. With the arrival of pre-trained models such as BERT, fine-tuning pre-trained models for downstream tasks became the norm. The increasing capabilities of ever larger models then enabled in-context learning via prompting. Recently, instruction tuning has become the newe&#8230;&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;&#129489;&#8205;&#127979; Instruction Tuning Vol. 1&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:7965403,&quot;name&quot;:&quot;Sebastian Ruder&quot;,&quot;bio&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/17fdd4c3-a575-4fe4-b58e-d876b78bfe2f_2416x2416.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2023-10-04T08:00:22.414Z&quot;,&quot;cover_image&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://nlpnewsletter.substack.com/p/instruction-tuning-vol-1&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:136684903,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:44,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;NLP News&quot;,&quot;publication_logo_url&quot;:&quot;https://s3.amazonaws.com/revue/profiles/images/000/027/670/thumb/sebastian_ruder_profile_photo_square.jpg?1616408613&quot;,&quot;belowTheFold&quot;:false,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><p>This month, we cover the latest datasets that are now much closer to real-world use cases&#8212;but still have their limitations! </p><div><hr></div><h2>Initiatives to Get Involved in AI Research</h2><p>Before we get into these, here are a few initiatives to get involved in AI research:</p><ul><li><p><strong><a href="https://aya.for.ai/">The Aya Project</a>:</strong> Help build state-of-the-art multilingual LLMs and make LLMs accessible in your language. <a href="https://txt.cohere.com/aya-multilingual/">Aya launched in January 2023</a> and is now in the final stage of dataset creation, so now is the time to <a href="https://sites.google.com/corp/cohere.com/aya-en/home?authuser=0">get involved</a> to make a difference!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yFm4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48f89242-521e-4e54-9cf4-5c42e9da21af_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yFm4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48f89242-521e-4e54-9cf4-5c42e9da21af_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!yFm4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48f89242-521e-4e54-9cf4-5c42e9da21af_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!yFm4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48f89242-521e-4e54-9cf4-5c42e9da21af_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!yFm4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48f89242-521e-4e54-9cf4-5c42e9da21af_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yFm4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48f89242-521e-4e54-9cf4-5c42e9da21af_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/48f89242-521e-4e54-9cf4-5c42e9da21af_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:40009,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yFm4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48f89242-521e-4e54-9cf4-5c42e9da21af_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!yFm4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48f89242-521e-4e54-9cf4-5c42e9da21af_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!yFm4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48f89242-521e-4e54-9cf4-5c42e9da21af_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!yFm4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48f89242-521e-4e54-9cf4-5c42e9da21af_960x540.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div></li><li><p><strong><a href="https://github.com/SEACrowd">SEACrowd</a>:</strong> Make NLP and LLM models more accessible in South-East Asian (SEA) languages by <a href="https://x.com/AlhamFikri/status/1719263648382853380?s=20">gathering and standardizing SEA datasets</a>. Contributions earn merch &#128085; and co-authorship &#128221;!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zzUt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80439b83-b147-413b-a9bc-e861f2cc6cdf_1920x1080.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zzUt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80439b83-b147-413b-a9bc-e861f2cc6cdf_1920x1080.jpeg 424w, https://substackcdn.com/image/fetch/$s_!zzUt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80439b83-b147-413b-a9bc-e861f2cc6cdf_1920x1080.jpeg 848w, https://substackcdn.com/image/fetch/$s_!zzUt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80439b83-b147-413b-a9bc-e861f2cc6cdf_1920x1080.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!zzUt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80439b83-b147-413b-a9bc-e861f2cc6cdf_1920x1080.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zzUt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80439b83-b147-413b-a9bc-e861f2cc6cdf_1920x1080.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/80439b83-b147-413b-a9bc-e861f2cc6cdf_1920x1080.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:271310,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zzUt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80439b83-b147-413b-a9bc-e861f2cc6cdf_1920x1080.jpeg 424w, https://substackcdn.com/image/fetch/$s_!zzUt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80439b83-b147-413b-a9bc-e861f2cc6cdf_1920x1080.jpeg 848w, https://substackcdn.com/image/fetch/$s_!zzUt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80439b83-b147-413b-a9bc-e861f2cc6cdf_1920x1080.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!zzUt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80439b83-b147-413b-a9bc-e861f2cc6cdf_1920x1080.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div></li><li><p><strong><a href="https://tropical.probabilistic.ai/">Tropical ProbAI</a>:</strong> Applications are open for the first Tropical Probabilistic AI School from 29 January &#8211; 2 February 2024 in Rio de Janeiro, Brazil. <a href="https://tropical.probabilistic.ai/application/">Application deadline is November 24</a>. Tropical ProbAI is also <a href="https://tropical.probabilistic.ai/#sponsors">looking for sponsors</a> to help bring AI schools to the under-served Latin American community</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_fS4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94310b8c-01a4-40e5-bfe0-4a69942a284f_2314x1174.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_fS4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94310b8c-01a4-40e5-bfe0-4a69942a284f_2314x1174.png 424w, https://substackcdn.com/image/fetch/$s_!_fS4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94310b8c-01a4-40e5-bfe0-4a69942a284f_2314x1174.png 848w, https://substackcdn.com/image/fetch/$s_!_fS4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94310b8c-01a4-40e5-bfe0-4a69942a284f_2314x1174.png 1272w, https://substackcdn.com/image/fetch/$s_!_fS4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94310b8c-01a4-40e5-bfe0-4a69942a284f_2314x1174.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_fS4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94310b8c-01a4-40e5-bfe0-4a69942a284f_2314x1174.png" width="1456" height="739" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/94310b8c-01a4-40e5-bfe0-4a69942a284f_2314x1174.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:739,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3130973,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_fS4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94310b8c-01a4-40e5-bfe0-4a69942a284f_2314x1174.png 424w, https://substackcdn.com/image/fetch/$s_!_fS4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94310b8c-01a4-40e5-bfe0-4a69942a284f_2314x1174.png 848w, https://substackcdn.com/image/fetch/$s_!_fS4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94310b8c-01a4-40e5-bfe0-4a69942a284f_2314x1174.png 1272w, https://substackcdn.com/image/fetch/$s_!_fS4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94310b8c-01a4-40e5-bfe0-4a69942a284f_2314x1174.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div></li></ul><p><em>Are there any other projects or initiatives that people should be aware of? Let me know.</em></p><div><hr></div><h2>Characteristics of Instruction-Tuning Data</h2><p>Now, let&#8217;s get on with instruction tuning. There are a few things to consider when using these recent datasets:</p><ol><li><p><strong>Data source:</strong> How was the data obtained? Most datasets have been generated using ChatGPT. They may thus inherit biases of the source model or may be noisy. Human-written examples are more expensive to obtain but are more high quality.</p></li><li><p><strong>Data quality:</strong> Was any filtering done to improve the quality of the generated data? In most cases, filtering is based on simple heuristics or a pre-trained model, which can result in noisy data. The authors of <a href="https://arxiv.org/abs/2304.07327">OpenAssistant Conversations</a> went the extra mile and obtained human-annotated data quality labels. &#128170;</p></li><li><p><strong>Domain and language coverage:</strong> Most datasets cover general QA-style use cases and are in English. However, similar methods can be used to obtain data in other domains or languages.</p></li><li><p><strong>Number of dialog turns:</strong> A <a href="https://en.wikipedia.org/wiki/Turn-taking">dialog turn</a> is an utterance by one speaker. Most datasets are single-turn, i.e., they consist of a prompt and a single response. Multi-turn data may be necessary to train a more conversational model.</p></li><li><p><strong>License terms:</strong> Data generated using OpenAI models is subject to the <a href="https://openai.com/policies/terms-of-use">OpenAI terms of use</a>, which prohibit using the data to develop competing models. So look for data with a more permissive license to avoid any legal complications.</p></li></ol><div><hr></div><h2>The Latest Instruction-Tuning Datasets</h2><p>Let&#8217;s now take a look at the latest generation of instruction-tuning datasets:</p><ol><li><p><strong><a href="https://github.com/tatsu-lab/stanford_alpaca#data-release">Alpaca data</a> <a href="https://crfm.stanford.edu/2023/03/13/alpaca.html">(Taori et al., March 2023)</a></strong>: 52k English instruction examples generated using OpenAI&#8217;s text-davinci-003 with self-instruct (see the previous post for a discussion): </p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;aead6bbb-af62-4047-a04e-fca1ce394575&quot;,&quot;caption&quot;:&quot;NLP and ML have gone through several phases of how models are trained in recent years. With the arrival of pre-trained models such as BERT, fine-tuning pre-trained models for downstream tasks became the norm. The increasing capabilities of ever larger models then enabled in-context learning via prompting. Recently, instruction tuning has become the newe&#8230;&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;&#129489;&#8205;&#127979; Instruction Tuning Vol. 1&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:7965403,&quot;name&quot;:&quot;Sebastian Ruder&quot;,&quot;bio&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/17fdd4c3-a575-4fe4-b58e-d876b78bfe2f_2416x2416.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2023-10-04T08:00:22.414Z&quot;,&quot;cover_image&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://nlpnewsletter.substack.com/p/instruction-tuning-vol-1&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:136684903,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:29,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;NLP News&quot;,&quot;publication_logo_url&quot;:&quot;https://s3.amazonaws.com/revue/profiles/images/000/027/670/thumb/sebastian_ruder_profile_photo_square.jpg?1616408613&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><p>The authors applied <a href="https://github.com/tatsu-lab/stanford_alpaca#data-generation-process">some modifications</a> to simplify the data generation pipeline and lower costs&#8212;the final data cost less than $500 to generate! </p></li><li><p><strong>Evol-Instruct <a href="https://arxiv.org/abs/2304.12244">(Xu et al., April 2023)</a></strong>: A rewritten set of 250k English instruction-response pairs based on the Alpaca data. Instructions are rewritten a) to make them more complex or b) to create a new, more specialized instruction by prompting ChatGPT. In a second step, ChatGPT is used to generate the corresponding responses. Low-quality instruction-response pairs are filtered using heuristics. This process is repeated three times.</p></li><li><p><strong>Vicuna ShareGPT data <a href="https://lmsys.org/blog/2023-03-30-vicuna/">(Chiang et al., March 2023)</a></strong>: 70k English conversations shared by users and scraped from <a href="https://sharegpt.com/">sharegpt.com</a>. Pre-processing involved converting HTML to markdown, filtering out low-quality samples, and splitting lengthy conversations into smaller segments. Compared to the above single-turn datasets, the ShareGPT conversations often consist of multiple turns and are thus more useful for training a model to leverage the context of the conversation. The conversations may be owned by the users so their use is potentially problematic.</p></li><li><p><strong><a href="https://github.com/project-baize/baize-chatbot">Baize data</a> <a href="https://arxiv.org/abs/2304.01196">(Xu et al., April 2023)</a>: </strong>54k and 57k English multi-turn dialog examples (3.4 turns on average) generated with ChatGPT using questions from Quora and StackOverflow datasets respectively as seeds. ChatGPT simulates both the human and AI participants of the conversation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> In addition, they also generated 47k dialogs in the medical domain based on <a href="https://github.com/abachaa/MedQuAD">MedQuAD</a> questions. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rVm5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9522f3f-b75b-4e4d-bf04-5999d9a56a35_1840x842.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rVm5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9522f3f-b75b-4e4d-bf04-5999d9a56a35_1840x842.png 424w, https://substackcdn.com/image/fetch/$s_!rVm5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9522f3f-b75b-4e4d-bf04-5999d9a56a35_1840x842.png 848w, https://substackcdn.com/image/fetch/$s_!rVm5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9522f3f-b75b-4e4d-bf04-5999d9a56a35_1840x842.png 1272w, https://substackcdn.com/image/fetch/$s_!rVm5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9522f3f-b75b-4e4d-bf04-5999d9a56a35_1840x842.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rVm5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9522f3f-b75b-4e4d-bf04-5999d9a56a35_1840x842.png" width="1456" height="666" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c9522f3f-b75b-4e4d-bf04-5999d9a56a35_1840x842.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:666,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:244386,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rVm5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9522f3f-b75b-4e4d-bf04-5999d9a56a35_1840x842.png 424w, https://substackcdn.com/image/fetch/$s_!rVm5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9522f3f-b75b-4e4d-bf04-5999d9a56a35_1840x842.png 848w, https://substackcdn.com/image/fetch/$s_!rVm5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9522f3f-b75b-4e4d-bf04-5999d9a56a35_1840x842.png 1272w, https://substackcdn.com/image/fetch/$s_!rVm5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9522f3f-b75b-4e4d-bf04-5999d9a56a35_1840x842.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Multi-turn dialogue data generated using ChatGPT based on a seed from Quora <a href="https://arxiv.org/abs/2304.01196">(Xu et al., 2023)</a>.</figcaption></figure></div></li><li><p><strong><a href="https://huggingface.co/datasets/databricks/databricks-dolly-15k">databricks-dolly-15k</a> (<a href="https://www.databricks.com/blog/2023/04/12/dolly-first-open-commercially-viable-instruction-tuned-llm">Conover et al., April 2023</a>):</strong> 15k English instruction-following examples written by Databricks employees. Crucially, both instructions and answers are human-generated. This in contrast to the other datasets above where instruction and/or answers are generated by ChatGPT. Examples cover 7 use cases: open QA, closed QA, information extraction and summarization of Wikipedia data, brainstorming, classification, and creative writing. Compared to the other datasets above, the data is released under a permissive license that also allows for commercial use.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!F3Pp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8931029a-4520-4f88-9102-130aec82ae51_2094x916.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!F3Pp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8931029a-4520-4f88-9102-130aec82ae51_2094x916.png 424w, https://substackcdn.com/image/fetch/$s_!F3Pp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8931029a-4520-4f88-9102-130aec82ae51_2094x916.png 848w, https://substackcdn.com/image/fetch/$s_!F3Pp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8931029a-4520-4f88-9102-130aec82ae51_2094x916.png 1272w, https://substackcdn.com/image/fetch/$s_!F3Pp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8931029a-4520-4f88-9102-130aec82ae51_2094x916.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!F3Pp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8931029a-4520-4f88-9102-130aec82ae51_2094x916.png" width="1456" height="637" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8931029a-4520-4f88-9102-130aec82ae51_2094x916.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:637,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:332188,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!F3Pp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8931029a-4520-4f88-9102-130aec82ae51_2094x916.png 424w, https://substackcdn.com/image/fetch/$s_!F3Pp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8931029a-4520-4f88-9102-130aec82ae51_2094x916.png 848w, https://substackcdn.com/image/fetch/$s_!F3Pp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8931029a-4520-4f88-9102-130aec82ae51_2094x916.png 1272w, https://substackcdn.com/image/fetch/$s_!F3Pp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8931029a-4520-4f88-9102-130aec82ae51_2094x916.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">An Open QA example in <code>databricks-dolly-15k</code>.</figcaption></figure></div></li><li><p><strong><a href="https://huggingface.co/datasets/OpenAssistant/oasst1">OpenAssistant Conversations</a> (<a href="https://arxiv.org/abs/2304.07327">K&#246;pf et al., April 2023</a>):</strong> 11k crowd-sourced multilingual instruction-following conversations (for 52k examples, only the prompts are available). Human annotators generated messages for both the assistant and the human participant. The data differs in several aspects from the other datasets: <strong>1)</strong> it is multilingual (42.8% examples are in English, 31.4% in Spanish, and the rest in other languages); <strong>2)</strong> annotators annotated the quality of prompts and responses (460k quality ratings overall); and <strong>3)</strong> the annotators were provided with detailed guidelines, both for writing prompts and for acting as the assistant. The data uses a permissive license, which allows commercial use.</p></li><li><p><strong>LIMA data <a href="https://arxiv.org/abs/2305.11206">(Zhou et al., May 2023)</a>:</strong> 1k training and 300 test answer&#8211;response pairs mostly sampled from StackExchange, wikiHow and the Pushshift Reddit dataset with around 400 written by the paper authors. A nice observation of this study is that <strong>training on this small set of curated instruction data outperforms training on the much larger, noisier Alpaca data</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pA38!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pA38!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png 424w, https://substackcdn.com/image/fetch/$s_!pA38!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png 848w, https://substackcdn.com/image/fetch/$s_!pA38!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png 1272w, https://substackcdn.com/image/fetch/$s_!pA38!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pA38!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png" width="1456" height="893" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:893,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:405753,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pA38!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png 424w, https://substackcdn.com/image/fetch/$s_!pA38!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png 848w, https://substackcdn.com/image/fetch/$s_!pA38!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png 1272w, https://substackcdn.com/image/fetch/$s_!pA38!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64293fa7-f3f0-4cd6-80e5-01cdad4493b4_1597x980.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Examples from the manually authored portion of the LIMA dataset.</figcaption></figure></div></li></ol><h2>Takeaways</h2><p><strong>&#9989; Quality &gt; quantity.  </strong>As <a href="https://arxiv.org/abs/2305.11206">Zhou et al. (2023)</a> observe, training on a small set of high-quality data outperforms instruction-tuning on larger, noisier data. Using more diverse prompts and quality filtering both improve performance.</p><p>&#129489;&#8205;&#127891; <strong>Imitation != mastery.</strong>  Models that are instruction-tuned on ChatGPT-generated data mimic ChatGPT&#8217;s style (and may thus fool human raters!) but not its factuality (<a href="https://arxiv.org/abs/2305.15717">Gudibande et al., May 2023</a>). They perform worse on standard benchmarks. Using stronger base models is the best way to address this. </p><p>&#127963;&#65039; <strong>The stronger the base, the better.</strong>  More powerful base models also produce stronger instruction-tuned models <a href="https://arxiv.org/abs/2306.04751">(Wang et al., June 2023)</a>.</p><p>&#129351; <strong>The combination wins.</strong>  Combining multiple instruction-tuning datasets results in the best average performance across tasks <a href="https://arxiv.org/abs/2306.04751">(Wang et al., June 2023)</a>. Dataset mixing and developing <a href="https://www.modulardeeplearning.com/">modular instruction-tuned models</a> are thus important research directions.</p><h2>Future Directions</h2><p><strong>Understanding instruction-tuning.</strong>  While we have seen a proliferation of instruction-tuning datasets, we still lack a clear understanding of what makes a good instruction and good instruction&#8211;response pairs. There is much anecdotal knowledge when it comes to creating good model prompts&#8212;but to my knowledge it is unclear how instruction&#8211;following data can be created at scale in a more principled manner.</p><p><strong>Improving data quality.</strong>  To improve model performance, we need to develop more reliable methods to identify high-quality examples and filter out undesirable ones. In a similar vein, it is important to develop methods that allow us to identify how a particular instance affects model behavior and alignment at test time.</p><p><strong>Evaluating instruction-tuned models.</strong>  In light of the biases of both human and automatic evaluations, there is no clear gold standard for how to evaluate instruction-tuned models. Evaluating a model on a set of tests that can be efficiently and automatically evaluated is one way to side-step this issue, see LMentry (<a href="https://aclanthology.org/2023.findings-acl.666/">Efrat et al., ACL 2023</a>), M2C (<a href="https://aclanthology.org/2023.acl-long.396/">Hlavnova &amp; Ruder, ACL 2023</a>), IFEval (<a href="https://arxiv.org/abs/2311.07911">Zhou et al., Nov 2023</a>), etc but these are restricted to a certain set of use cases. In general, it is crucial to design evaluations with a target application in mind.</p><p><em>Are there any exciting developments or directions that I missed? Let me know.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">NLP News is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p><a href="https://arxiv.org/abs/2304.01196">Xu et al. (2023)</a> refer to this setting as &#8216;self-chat&#8217;. This is a continuation of work on machine-to-machine dialogue modeling (see, e.g., <a href="https://aclanthology.org/D18-1547/">Budzianowski et al., 2018</a>) and dialogue self-play (<a href="https://aclanthology.org/N18-3006/">Shah et al., 2018</a>).</p></div></div>]]></content:encoded></item><item><title><![CDATA[🌍⏳ Do LMs Represent Space and Time?]]></title><description><![CDATA[In this post, we&#8217;ll take a closer look at the question &#8220;Do LMs represent space and time?&#8221; inspired by a recent paper.]]></description><link>https://newsletter.ruder.io/p/do-lms-represent-space-and-time</link><guid isPermaLink="false">https://newsletter.ruder.io/p/do-lms-represent-space-and-time</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Mon, 09 Oct 2023 08:25:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xaEE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In this post, we&#8217;ll take a closer look at the question &#8220;Do LMs represent space and time?&#8221; inspired by a recent paper. We&#8217;ll look at how spatial and temporal information has been encoded in LMs, what this means for practical applications, as well as other aspects such as encoding of fine-grained spatial information and across cultures.</p><div><hr></div><h4><strong>Language Models Represent Space and Time (Gurnee &amp; Tegmark, Oct &#8216;23)</strong></h4><p>In a <a href="https://arxiv.org/abs/2310.02207">recent paper</a>, Gurnee and Tegmark show that LLMs (Llama-2 models specifically) learn linear representations of space and time. <strong>What does this mean exactly?</strong></p><p>The general setup looks like this:</p><ol><li><p>The authors process the names of places and historical figures with Llama-2. They create their own dataset, sourced from Wikipedia, for this.</p></li><li><p>They then take the hidden state of the last token of the entity (for each layer) as a representation of the entity name.</p></li><li><p>Finally, they train a linear probe (a one-layer MLP) on the representation to predict its coordinates (latitude and longitude) or year of death.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p></li></ol><p>They find that spatial and temporal information in Llama-2 can indeed be recovered with a linear probe, larger models are better at encoding this information, and that representations in the upper layers (from the middle layer to the last) achieve the highest accuracy. In other words, <strong>models learn a representation of places that&#8212;after a linear transformation&#8212;is more or less consistent with their location on a map.</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xaEE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xaEE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png 424w, https://substackcdn.com/image/fetch/$s_!xaEE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png 848w, https://substackcdn.com/image/fetch/$s_!xaEE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png 1272w, https://substackcdn.com/image/fetch/$s_!xaEE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xaEE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png" width="1148" height="499" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:499,&quot;width&quot;:1148,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:486157,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xaEE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png 424w, https://substackcdn.com/image/fetch/$s_!xaEE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png 848w, https://substackcdn.com/image/fetch/$s_!xaEE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png 1272w, https://substackcdn.com/image/fetch/$s_!xaEE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe690953d-a135-40b5-95ba-0e2cc616466e_1148x499.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Coordinates of place names of different continents predicted by a linear probe on Llama-2-70B&#8217;s layer 50. Predicted locations are close to their actual locations (R&#178;=0.92).</figcaption></figure></div><p>That&#8217;s a pretty cool finding and visualization. However, how surprising is it that current LLMs encode a map-like representation of places? On the whole, not very.</p><h4>Spatial Relations in word2vec</h4><p>Alexander Doria already <a href="https://x.com/Dorialexander/status/1709685322857042145?s=20">pointed out on Twitter</a> that this is not a new observation and that geographic relationships have been encoded already by much older models. The classic example is the word analogy task for word embeddings, which identifies representations in the embedding space using simple vector offset such as<br><code>Paris - France + Italy = ? </code> where the answer (i.e., the closest nearest neighbor in the embedding space) is expected to be Rome in this case.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> In the embedding literature, models generally performed very well at encoding such semantic relations.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cEb5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe11e57c-27f1-4dcd-85b2-5a3003a3f548_1492x843.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cEb5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe11e57c-27f1-4dcd-85b2-5a3003a3f548_1492x843.png 424w, https://substackcdn.com/image/fetch/$s_!cEb5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe11e57c-27f1-4dcd-85b2-5a3003a3f548_1492x843.png 848w, https://substackcdn.com/image/fetch/$s_!cEb5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe11e57c-27f1-4dcd-85b2-5a3003a3f548_1492x843.png 1272w, https://substackcdn.com/image/fetch/$s_!cEb5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe11e57c-27f1-4dcd-85b2-5a3003a3f548_1492x843.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cEb5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe11e57c-27f1-4dcd-85b2-5a3003a3f548_1492x843.png" width="1456" height="823" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/be11e57c-27f1-4dcd-85b2-5a3003a3f548_1492x843.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:823,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:128521,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cEb5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe11e57c-27f1-4dcd-85b2-5a3003a3f548_1492x843.png 424w, https://substackcdn.com/image/fetch/$s_!cEb5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe11e57c-27f1-4dcd-85b2-5a3003a3f548_1492x843.png 848w, https://substackcdn.com/image/fetch/$s_!cEb5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe11e57c-27f1-4dcd-85b2-5a3003a3f548_1492x843.png 1272w, https://substackcdn.com/image/fetch/$s_!cEb5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe11e57c-27f1-4dcd-85b2-5a3003a3f548_1492x843.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Skip-gram vectors of countries and corresponding capitals projected by PCA show their linear relationship <a href="https://arxiv.org/abs/1310.4546">(Mikolov et al., 2013)</a>.</figcaption></figure></div><p>These well-known relationships are between countries and their capitals. <strong>What about more fine-grained information such as the spatial coordinates investigated by Gurnee and Tegmark?</strong></p><h4>Analyzing Geographic Knowledge</h4><p>There is an existing thread of research that has focused on injecting and analyzing geographic knowledge in models. <a href="https://aclanthology.org/D18-1469">Hovy and Purschke (2018)</a> learn continuous representations of German cities using doc2vec. More recently, <a href="https://arxiv.org/abs/2203.08565">Hofmann et al. (2022)</a> adapt pre-trained BERT models in different languages with geographic knowledge by predicting geolocations on geo-labeled data while <a href="https://arxiv.org/abs/2212.10408">Faisal and Anastasopoulos (2022)</a> probe the geographic knowledge in GPT-2, mGPT, and BLOOM. In these studies, models do well at predicting place coordinates and produce map-like representations.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OYfR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc60fa495-9bfd-41c9-8058-39035906417d_1414x580.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OYfR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc60fa495-9bfd-41c9-8058-39035906417d_1414x580.png 424w, https://substackcdn.com/image/fetch/$s_!OYfR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc60fa495-9bfd-41c9-8058-39035906417d_1414x580.png 848w, https://substackcdn.com/image/fetch/$s_!OYfR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc60fa495-9bfd-41c9-8058-39035906417d_1414x580.png 1272w, https://substackcdn.com/image/fetch/$s_!OYfR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc60fa495-9bfd-41c9-8058-39035906417d_1414x580.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OYfR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc60fa495-9bfd-41c9-8058-39035906417d_1414x580.png" width="1414" height="580" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c60fa495-9bfd-41c9-8058-39035906417d_1414x580.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:580,&quot;width&quot;:1414,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:848293,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!OYfR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc60fa495-9bfd-41c9-8058-39035906417d_1414x580.png 424w, https://substackcdn.com/image/fetch/$s_!OYfR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc60fa495-9bfd-41c9-8058-39035906417d_1414x580.png 848w, https://substackcdn.com/image/fetch/$s_!OYfR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc60fa495-9bfd-41c9-8058-39035906417d_1414x580.png 1272w, https://substackcdn.com/image/fetch/$s_!OYfR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc60fa495-9bfd-41c9-8058-39035906417d_1414x580.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Clustered city embeddings with 3, 5, and 8 clusters <a href="https://aclanthology.org/D18-1469/">(Hovy &amp; Purschke, 2018)</a>.</figcaption></figure></div><p>There is other work that focuses specifically on the task of geolocation where even simple one-hidden layer MLPs can do well <a href="https://aclanthology.org/P17-2033/">(Rahimi et al., 2017)</a>. In light of this prior work, it is unsurprising that the latest LLMs encode spatial information. In addition, <strong>encoding spatial information does not seem to be a property that emerges only with sufficient model size</strong>. In order to know whether recent LLMs are actually more spatially aware than prior models, it is thus important to compare them to prior models and on established tasks such as user geolocation.</p><p>Overall, studies such as the one by Gurnee and Tegmark are crucial to get a better understanding of LLMs. However, rather than focusing solely on work on LLMs, these studies would benefit from being aware of and leveraging prior work as a source of baselines, evaluation datasets, and methods.</p><h4>LLMs as Geographic Information Systems</h4><p>As LLMs capture a surprising amount of geographic information, they may be useful for a range of other geography-related applications such as a <a href="https://en.wikipedia.org/wiki/Geographic_information_system">geographic information system (GIS)</a>. A GIS is a computer system that stores, checks, analyzes, and displays geographic data&#8212;something that could be emulated by an LLM. <a href="https://arxiv.org/abs/2305.06453">Li and Ning (May 2023)</a> show through GIS case studies how LLMs can be used, for instance, to identify the population living close to hazardous waste facilities and to map their distribution, among other applications.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!I551!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faed2eee3-f4eb-4eb1-81f6-623205bdfb48_1308x738.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!I551!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faed2eee3-f4eb-4eb1-81f6-623205bdfb48_1308x738.png 424w, https://substackcdn.com/image/fetch/$s_!I551!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faed2eee3-f4eb-4eb1-81f6-623205bdfb48_1308x738.png 848w, https://substackcdn.com/image/fetch/$s_!I551!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faed2eee3-f4eb-4eb1-81f6-623205bdfb48_1308x738.png 1272w, https://substackcdn.com/image/fetch/$s_!I551!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faed2eee3-f4eb-4eb1-81f6-623205bdfb48_1308x738.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!I551!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faed2eee3-f4eb-4eb1-81f6-623205bdfb48_1308x738.png" width="1308" height="738" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aed2eee3-f4eb-4eb1-81f6-623205bdfb48_1308x738.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:738,&quot;width&quot;:1308,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:401134,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!I551!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faed2eee3-f4eb-4eb1-81f6-623205bdfb48_1308x738.png 424w, https://substackcdn.com/image/fetch/$s_!I551!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faed2eee3-f4eb-4eb1-81f6-623205bdfb48_1308x738.png 848w, https://substackcdn.com/image/fetch/$s_!I551!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faed2eee3-f4eb-4eb1-81f6-623205bdfb48_1308x738.png 1272w, https://substackcdn.com/image/fetch/$s_!I551!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faed2eee3-f4eb-4eb1-81f6-623205bdfb48_1308x738.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Results generated by ChatGPT for counting the population living near hazardous waste facilities. (a) Solution graph, (b) Python code, and (c) returned population count and generated map <a href="https://arxiv.org/abs/2305.06453">(Li &amp; Ning, 2023)</a>.</figcaption></figure></div><p>Beyond accurately encoding and reasoning with geographic data, the use of LLMs as GIS thus also requires LLMs to interface with auxiliary tools such as data readers, calculators, code execution, and visualization, which I covered in a previous newsletter:</p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;f8417ad1-1254-4933-96b3-82c2ff98648b&quot;,&quot;caption&quot;:&quot;Hi all, It&#8217;s great to get back to writing regularly. Writing a newsletter provides me with an opportunity to delve into topics I&#8217;m excited about. In this edition, we&#8217;ll explore tool use&#8212;arguably one of the hottest new capabilities of LLMs. We&#8217;ll look at types of tools, benefits of tool use, recent developments, and future directions.&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;&#128736; Tool-Augmented LLMs&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:7965403,&quot;name&quot;:&quot;Sebastian Ruder&quot;,&quot;bio&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/17fdd4c3-a575-4fe4-b58e-d876b78bfe2f_2416x2416.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2023-08-28T09:51:45.517Z&quot;,&quot;cover_image&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://nlpnewsletter.substack.com/p/tool-augmented-llms&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:136248186,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:46,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;NLP News&quot;,&quot;publication_logo_url&quot;:&quot;https://s3.amazonaws.com/revue/profiles/images/000/027/670/thumb/sebastian_ruder_profile_photo_square.jpg?1616408613&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><h4>Encoding Fine-grained Spatial Relations</h4><p>LLMs encode spatial information on a <em>macro level</em>&#8212;related to cities and places&#8212;but what about<em> fine-grained spatial relations</em> such as whether something is behind, next to, left to, above, etc.? Are these also encoded in a consistent manner?</p><p>Prior work in this area <a href="https://arxiv.org/abs/1807.01670">(Ramalho et al., 2018)</a> generated images from their textual descriptions using a VAE to learn to encode spatial relations from natural language. Recently, <a href="https://arxiv.org/abs/2307.03678">Ji and Gao (July 2023)</a> evaluated GPT-2 and BERT on their ability to encode geometric attributes, achieving up to 73% accuracy on spatial relations. For the largest LLMs, I am only aware of case studies that show awareness of certain spatial relations such as <a href="https://arxiv.org/abs/2303.12712">Bubeck et al. (2023)</a> so there is potential for more work in this area. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8qwZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd2929c-09ce-4eb2-938b-fad07584087c_1443x388.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8qwZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd2929c-09ce-4eb2-938b-fad07584087c_1443x388.png 424w, https://substackcdn.com/image/fetch/$s_!8qwZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd2929c-09ce-4eb2-938b-fad07584087c_1443x388.png 848w, https://substackcdn.com/image/fetch/$s_!8qwZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd2929c-09ce-4eb2-938b-fad07584087c_1443x388.png 1272w, https://substackcdn.com/image/fetch/$s_!8qwZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd2929c-09ce-4eb2-938b-fad07584087c_1443x388.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8qwZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd2929c-09ce-4eb2-938b-fad07584087c_1443x388.png" width="1443" height="388" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dbd2929c-09ce-4eb2-938b-fad07584087c_1443x388.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:388,&quot;width&quot;:1443,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:74811,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8qwZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd2929c-09ce-4eb2-938b-fad07584087c_1443x388.png 424w, https://substackcdn.com/image/fetch/$s_!8qwZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd2929c-09ce-4eb2-938b-fad07584087c_1443x388.png 848w, https://substackcdn.com/image/fetch/$s_!8qwZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd2929c-09ce-4eb2-938b-fad07584087c_1443x388.png 1272w, https://substackcdn.com/image/fetch/$s_!8qwZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbd2929c-09ce-4eb2-938b-fad07584087c_1443x388.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">GPT-4 navigates a map interactively. Left: True map and exploration path of GPT-4. Right: The map generated by GPT-4 based on the prompt &#8220;<code>Can you draw a pyplot showing the position of the rooms/places and connect them using lines?</code>&#8221;. GPT-4 is able to track the locations and visualize them correctly <a href="https://arxiv.org/abs/2303.12712">(Bubeck et al., 2023)</a>.</figcaption></figure></div><h4>Encoding Time</h4><p>Regarding the encoding of time, it is important to look beyond synthetic tasks and to practical applications for evaluation. Given that the world we live in is constantly changing, it is critical to ensure that models reflect up-to-date information about the world. Prior work has used language modeling <a href="https://arxiv.org/abs/2102.01951">(Lazaridou et al., 2021)</a> and question answering <a href="https://aclanthology.org/2021.emnlp-main.586/">(Zhang &amp; Choi, 2021)</a> for model evaluation.</p><p>More recently, <a href="https://aclanthology.org/2023.acl-long.828/">Tan et al. (ACL 2023)</a> introduced a new temporal reasoning QA benchmark that assesses models on three levels of temporal reasoning: 1) relations between different times; 2) relations between times and events; and 3) relations between different events. Particularly time-event and event-event reasoning is still challenging even for the latest LLMs.</p><h4>Encoding Space and Time Across Cultures</h4><p>The way spatial and temporal information is expressed differs across languages and cultures. In Swahili, time is based on sunset and sunrise rather than a.m. and p.m. For example, 11.30 am in standard time is 5.30 in the morning in Swahili time. For a recent paper <a href="https://aclanthology.org/2023.acl-long.396/">(Hlavnova &amp; Ruder, ACL 2023)</a>, we evaluated LLMs on different types of reasoning across languages and found that they did much worse on languages with different time expressions such as Swahili. Similarly, understanding of time expressions can also be evaluated based on models&#8217; ability to ground time expressions, i.e., to map culture-specific time expressions such as &#8220;morning&#8221; in English or &#8220;manh&#227;&#8221; in Portuguese to specific hours in the day <a href="https://aclanthology.org/2022.findings-acl.224/">(Shwartz, 2022)</a>.</p><p>For spatial information, datasets such as MarVL <a href="https://aclanthology.org/2021.emnlp-main.818/">(Liu et al., 2021)</a> and Crossmodal-3600 <a href="https://aclanthology.org/2022.emnlp-main.45/">(Thapliyal et al., 2022)</a> can be used to investigate models&#8217; visual perception across cultures&#8212;but I&#8217;m not aware of any datasets that enable an analysis of cross-cultural encoding of spatial information.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ps9m!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aa54832-618b-47f2-b1e2-3da74540a583_1257x489.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ps9m!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aa54832-618b-47f2-b1e2-3da74540a583_1257x489.png 424w, https://substackcdn.com/image/fetch/$s_!Ps9m!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aa54832-618b-47f2-b1e2-3da74540a583_1257x489.png 848w, https://substackcdn.com/image/fetch/$s_!Ps9m!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aa54832-618b-47f2-b1e2-3da74540a583_1257x489.png 1272w, https://substackcdn.com/image/fetch/$s_!Ps9m!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aa54832-618b-47f2-b1e2-3da74540a583_1257x489.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ps9m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aa54832-618b-47f2-b1e2-3da74540a583_1257x489.png" width="1257" height="489" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0aa54832-618b-47f2-b1e2-3da74540a583_1257x489.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:489,&quot;width&quot;:1257,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:469780,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ps9m!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aa54832-618b-47f2-b1e2-3da74540a583_1257x489.png 424w, https://substackcdn.com/image/fetch/$s_!Ps9m!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aa54832-618b-47f2-b1e2-3da74540a583_1257x489.png 848w, https://substackcdn.com/image/fetch/$s_!Ps9m!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aa54832-618b-47f2-b1e2-3da74540a583_1257x489.png 1272w, https://substackcdn.com/image/fetch/$s_!Ps9m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aa54832-618b-47f2-b1e2-3da74540a583_1257x489.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Captions in three different languages for an image in Crossmodal-3600. Captions are created by native speakers in each language and are thus free of translation artifacts.</figcaption></figure></div><p>I hope you found this short review of space and time representations in language models interesting. Did I miss any interesting work in this space? What are your favorite observations and insights about how LLMs encode information? Let me know in the comments.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Training a linear classifier on the hidden representations of a model (without further fine-tuning) is standard methodology and has been used extensively to analyze BERT (<a href="https://arxiv.org/abs/2002.12327">Rogers et al., 2020</a>). For more information on probing pre-trained models, check out <a href="https://people.cs.umass.edu/~miyyer/cs685/slides/probes.pdf">these slides</a> by Mohit Iyyer (Intro to NLP Spring 2023; based on slides from Tu Vu).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>While there are well-known issues with the analogy formulation (see, for instance, <a href="https://ojs.aaai.org/index.php/AAAI/article/view/17524">Garneau et al., 2021</a>), it can help illustrate relationships encoded in the embedding space.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[🧑‍🏫 Instruction Tuning Vol. 1]]></title><description><![CDATA[The most popular instruction datasets]]></description><link>https://newsletter.ruder.io/p/instruction-tuning-vol-1</link><guid isPermaLink="false">https://newsletter.ruder.io/p/instruction-tuning-vol-1</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Wed, 04 Oct 2023 08:00:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>NLP and ML have gone through several phases of how models are trained in recent years. With the arrival of pre-trained models such as BERT, fine-tuning pre-trained models for downstream tasks became the norm. The increasing capabilities of ever larger models then enabled in-context learning via prompting. Recently, instruction tuning has become the newest method to make LLMs useful in practice.</p><p>In this edition, we will cover some of the most popular datasets for instruction tuning. The next editions will cover the latest instruction datasets and instruction-tuned models.</p><div><hr></div><h3>What is Instruction Tuning?</h3><p>The main difference between instruction tuning and standard supervised fine-tuning lies in the data that the model is trained on. Whereas supervised fine-tuning trains models on <strong>input examples and their corresponding outputs</strong>, instruction tuning <strong>augments input-output examples with instructions</strong>, which enables instruction-tuned models to generalize more easily to new tasks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pza7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4935aa8-722e-4461-9e9e-88341b5558f6_1379x446.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pza7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4935aa8-722e-4461-9e9e-88341b5558f6_1379x446.png 424w, https://substackcdn.com/image/fetch/$s_!pza7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4935aa8-722e-4461-9e9e-88341b5558f6_1379x446.png 848w, https://substackcdn.com/image/fetch/$s_!pza7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4935aa8-722e-4461-9e9e-88341b5558f6_1379x446.png 1272w, https://substackcdn.com/image/fetch/$s_!pza7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4935aa8-722e-4461-9e9e-88341b5558f6_1379x446.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pza7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4935aa8-722e-4461-9e9e-88341b5558f6_1379x446.png" width="1379" height="446" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c4935aa8-722e-4461-9e9e-88341b5558f6_1379x446.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:446,&quot;width&quot;:1379,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:148770,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pza7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4935aa8-722e-4461-9e9e-88341b5558f6_1379x446.png 424w, https://substackcdn.com/image/fetch/$s_!pza7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4935aa8-722e-4461-9e9e-88341b5558f6_1379x446.png 848w, https://substackcdn.com/image/fetch/$s_!pza7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4935aa8-722e-4461-9e9e-88341b5558f6_1379x446.png 1272w, https://substackcdn.com/image/fetch/$s_!pza7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4935aa8-722e-4461-9e9e-88341b5558f6_1379x446.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Pretrain-finetuning vs prompting vs instruction tuning (<a href="https://openreview.net/forum?id=gEZrGCozdqR">Wei et al., 2022</a>).</figcaption></figure></div><p>Methods differ based on how the instruction tuning data is constructed. <a href="https://arxiv.org/abs/2308.10792">Zhang et al. (2023)</a> provide a good overview of existing instruction datasets. Existing datasets fall roughly into two main categories: a) instructions are added to existing NLP tasks; and b) data from (a) is used to condition a model to generate new instruction-input-output tuples. Let&#8217;s now look at some of the most popular instruction datasets:</p><ol><li><p><strong><a href="https://instructions.apps.allenai.org/">Natural Instructions</a> (<a href="https://aclanthology.org/2022.acl-long.244/">Mishra et al., 2022</a>)</strong>: 193k instruction-output examples sourced from 61 existing English NLP tasks. Crowd-sourcing instructions from each dataset are aligned to a common schema. <strong>Instructions are thus more structured</strong> compared to other datasets. <strong>Outputs are relatively short</strong>, however, which makes the data less useful for generating long-form content.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!O7tL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39e03bdc-0c37-45ef-8b03-114ed4d57187_1467x868.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!O7tL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39e03bdc-0c37-45ef-8b03-114ed4d57187_1467x868.png 424w, https://substackcdn.com/image/fetch/$s_!O7tL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39e03bdc-0c37-45ef-8b03-114ed4d57187_1467x868.png 848w, https://substackcdn.com/image/fetch/$s_!O7tL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39e03bdc-0c37-45ef-8b03-114ed4d57187_1467x868.png 1272w, https://substackcdn.com/image/fetch/$s_!O7tL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39e03bdc-0c37-45ef-8b03-114ed4d57187_1467x868.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!O7tL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39e03bdc-0c37-45ef-8b03-114ed4d57187_1467x868.png" width="1456" height="861" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/39e03bdc-0c37-45ef-8b03-114ed4d57187_1467x868.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:861,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:511992,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!O7tL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39e03bdc-0c37-45ef-8b03-114ed4d57187_1467x868.png 424w, https://substackcdn.com/image/fetch/$s_!O7tL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39e03bdc-0c37-45ef-8b03-114ed4d57187_1467x868.png 848w, https://substackcdn.com/image/fetch/$s_!O7tL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39e03bdc-0c37-45ef-8b03-114ed4d57187_1467x868.png 1272w, https://substackcdn.com/image/fetch/$s_!O7tL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39e03bdc-0c37-45ef-8b03-114ed4d57187_1467x868.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Two instances of Natural Instructions. The instruction schema covers multiple fields including a definition, things to avoid, and positive and negative examples.</figcaption></figure></div></li><li><p><strong>Natural Instructions v2 / Super-Natural Instructions (<a href="https://aclanthology.org/2022.emnlp-main.340/">Wang et al., 2022</a>)</strong>: A crowd-sourced collection of instruction data based on existing NLP tasks and simple synthetic tasks. It includes 5M examples across 76 tasks in 55 languages. Compared to Natural Instructions, <strong>instructions are simplified</strong>; they consist of a task definition and positive and negative examples with explanations.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!foDw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!foDw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png 424w, https://substackcdn.com/image/fetch/$s_!foDw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png 848w, https://substackcdn.com/image/fetch/$s_!foDw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png 1272w, https://substackcdn.com/image/fetch/$s_!foDw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!foDw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png" width="718" height="929" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:929,&quot;width&quot;:718,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:256374,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!foDw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png 424w, https://substackcdn.com/image/fetch/$s_!foDw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png 848w, https://substackcdn.com/image/fetch/$s_!foDw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png 1272w, https://substackcdn.com/image/fetch/$s_!foDw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d8e6cae-c3f0-4776-a355-a7d9b149b6bc_718x929.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">An instance of Super-Natural Instructions. Instructions consist of a task definition and positive and negative examples with an explanation.</figcaption></figure></div></li><li><p><strong>Unnatural Instructions (<a href="https://aclanthology.org/2023.acl-long.806/">Honovich et al., 2023</a>)</strong>: An automatically collected instruction dataset of 240k examples where InstructGPT (text-davinci-002) is prompted with three Super-Natural Instructions examples&#8212;consisting of an instruction, input, possible output constraints&#8212;and asked to generate a new example. The output is generated separately by conditioning on the generated instruction, input, and constraints. The generated instructions are then further paraphrased by prompting the model. Unnatural Instructions covers a <strong>more diverse set of tasks</strong> than Super-Natural Instructions; while many examples reflect classical NLP tasks, it also includes examples of other interesting tasks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Kz8N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb808851-1521-4293-932f-f5da0704227e_1447x902.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Kz8N!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb808851-1521-4293-932f-f5da0704227e_1447x902.png 424w, https://substackcdn.com/image/fetch/$s_!Kz8N!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb808851-1521-4293-932f-f5da0704227e_1447x902.png 848w, https://substackcdn.com/image/fetch/$s_!Kz8N!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb808851-1521-4293-932f-f5da0704227e_1447x902.png 1272w, https://substackcdn.com/image/fetch/$s_!Kz8N!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb808851-1521-4293-932f-f5da0704227e_1447x902.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Kz8N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb808851-1521-4293-932f-f5da0704227e_1447x902.png" width="1447" height="902" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cb808851-1521-4293-932f-f5da0704227e_1447x902.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:902,&quot;width&quot;:1447,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:345803,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Kz8N!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb808851-1521-4293-932f-f5da0704227e_1447x902.png 424w, https://substackcdn.com/image/fetch/$s_!Kz8N!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb808851-1521-4293-932f-f5da0704227e_1447x902.png 848w, https://substackcdn.com/image/fetch/$s_!Kz8N!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb808851-1521-4293-932f-f5da0704227e_1447x902.png 1272w, https://substackcdn.com/image/fetch/$s_!Kz8N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb808851-1521-4293-932f-f5da0704227e_1447x902.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Examples of eight interesting generated instructions that differ from classical NLP tasks in Unnatural Instructions.</figcaption></figure></div></li><li><p><strong>Self-Instruct <a href="https://aclanthology.org/2023.acl-long.754/">(Wang et al., 2023)</a></strong>: Similar to Unnatural Instructions, Self-Instruct consists of 82k examples automatically generated using InstructGPT conditioned on a set of <a href="https://github.com/yizhongw/self-instruct/blob/main/human_eval/user_oriented_instructions.jsonl">seed task examples</a> (175 tasks in total; one example per task; 8 examples are sampled for in-context learning). Self-Instruct decouples the example generation by first generating the instruction, then the input (conditioned on instruction), and then the output. For classification tasks, the authors first generate the possible output labels and then condition the input generation on each class label to avoid biasing towards a specific label. While the generated instructions are mostly valid, the <strong>generated outputs are often noisy</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pJBM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff0c024-6976-40c5-8438-39a963fcaf63_1538x399.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pJBM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff0c024-6976-40c5-8438-39a963fcaf63_1538x399.png 424w, https://substackcdn.com/image/fetch/$s_!pJBM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff0c024-6976-40c5-8438-39a963fcaf63_1538x399.png 848w, https://substackcdn.com/image/fetch/$s_!pJBM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff0c024-6976-40c5-8438-39a963fcaf63_1538x399.png 1272w, https://substackcdn.com/image/fetch/$s_!pJBM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff0c024-6976-40c5-8438-39a963fcaf63_1538x399.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pJBM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff0c024-6976-40c5-8438-39a963fcaf63_1538x399.png" width="1456" height="378" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9ff0c024-6976-40c5-8438-39a963fcaf63_1538x399.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:378,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:156098,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pJBM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff0c024-6976-40c5-8438-39a963fcaf63_1538x399.png 424w, https://substackcdn.com/image/fetch/$s_!pJBM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff0c024-6976-40c5-8438-39a963fcaf63_1538x399.png 848w, https://substackcdn.com/image/fetch/$s_!pJBM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff0c024-6976-40c5-8438-39a963fcaf63_1538x399.png 1272w, https://substackcdn.com/image/fetch/$s_!pJBM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff0c024-6976-40c5-8438-39a963fcaf63_1538x399.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Examples of generated instruction-input-output tuples in Self-Instruct.</figcaption></figure></div></li><li><p><strong><a href="https://huggingface.co/datasets/bigscience/P3">P3</a> (<a href="https://openreview.net/forum?id=9Vrb9D0WI4">Public Pool of Prompts; Sanh et al., 2022</a>)</strong>: A crowd-sourced collection of prompts for 177 English NLP tasks. For each dataset, about 11 different prompts are available on average, which enables studying the impact of different prompt formulations. Compared to the instructions in the above instruction datasets, <strong>P3 prompts are often shorter and less elaborate</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Efrh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F995dcc96-3ae0-4c56-9a09-0c7d642b0822_1339x451.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Efrh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F995dcc96-3ae0-4c56-9a09-0c7d642b0822_1339x451.png 424w, https://substackcdn.com/image/fetch/$s_!Efrh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F995dcc96-3ae0-4c56-9a09-0c7d642b0822_1339x451.png 848w, https://substackcdn.com/image/fetch/$s_!Efrh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F995dcc96-3ae0-4c56-9a09-0c7d642b0822_1339x451.png 1272w, https://substackcdn.com/image/fetch/$s_!Efrh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F995dcc96-3ae0-4c56-9a09-0c7d642b0822_1339x451.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Efrh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F995dcc96-3ae0-4c56-9a09-0c7d642b0822_1339x451.png" width="1339" height="451" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/995dcc96-3ae0-4c56-9a09-0c7d642b0822_1339x451.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:451,&quot;width&quot;:1339,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:130448,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Efrh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F995dcc96-3ae0-4c56-9a09-0c7d642b0822_1339x451.png 424w, https://substackcdn.com/image/fetch/$s_!Efrh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F995dcc96-3ae0-4c56-9a09-0c7d642b0822_1339x451.png 848w, https://substackcdn.com/image/fetch/$s_!Efrh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F995dcc96-3ae0-4c56-9a09-0c7d642b0822_1339x451.png 1272w, https://substackcdn.com/image/fetch/$s_!Efrh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F995dcc96-3ae0-4c56-9a09-0c7d642b0822_1339x451.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">P3 prompt templates for two existing NLP tasks. Prompt templates use fields of the raw data (e.g., <em>{Document}</em>) and template metadata (e.g., <em>{Choices[label]}</em>).</figcaption></figure></div></li><li><p><strong><a href="https://github.com/bigscience-workshop/xmtf">xP3, xP3mt</a> <a href="https://aclanthology.org/2023.acl-long.891/">(Muennighoff et al., 2023)</a></strong>: An extension of P3 including 19 multilingual datasets and 11 code datasets, with English prompts. They also release a machine-translated version of the data (xP3mt), which contains prompts automatically translated into 20 languages. Fine-tuning on multilingual tasks with English prompts further improves performance beyond only fine-tuning on English instruction data.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!opoo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2e8b350-d214-405f-9848-fca7e9a8b244_1470x403.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!opoo!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2e8b350-d214-405f-9848-fca7e9a8b244_1470x403.png 424w, https://substackcdn.com/image/fetch/$s_!opoo!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2e8b350-d214-405f-9848-fca7e9a8b244_1470x403.png 848w, https://substackcdn.com/image/fetch/$s_!opoo!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2e8b350-d214-405f-9848-fca7e9a8b244_1470x403.png 1272w, https://substackcdn.com/image/fetch/$s_!opoo!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2e8b350-d214-405f-9848-fca7e9a8b244_1470x403.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!opoo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2e8b350-d214-405f-9848-fca7e9a8b244_1470x403.png" width="1456" height="399" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c2e8b350-d214-405f-9848-fca7e9a8b244_1470x403.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:399,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:110244,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!opoo!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2e8b350-d214-405f-9848-fca7e9a8b244_1470x403.png 424w, https://substackcdn.com/image/fetch/$s_!opoo!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2e8b350-d214-405f-9848-fca7e9a8b244_1470x403.png 848w, https://substackcdn.com/image/fetch/$s_!opoo!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2e8b350-d214-405f-9848-fca7e9a8b244_1470x403.png 1272w, https://substackcdn.com/image/fetch/$s_!opoo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2e8b350-d214-405f-9848-fca7e9a8b244_1470x403.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Prompt templates in P3, xP3, and xP3mt. Prompts are in English in P3 and xP3. </figcaption></figure></div></li><li><p><strong>Flan 2021 / Muffin (<a href="https://openreview.net/forum?id=gEZrGCozdqR">Wei et al., 2022</a>): </strong>Prompts for 62 English text datasets, with 10 prompt templates for each task. For classification tasks, an <code>OPTIONS</code> suffix is appended to the input in order to indicate output constraints.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!B6fz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fceb4e28d-424d-450d-a29e-598d883d1fb2_1229x500.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!B6fz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fceb4e28d-424d-450d-a29e-598d883d1fb2_1229x500.png 424w, https://substackcdn.com/image/fetch/$s_!B6fz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fceb4e28d-424d-450d-a29e-598d883d1fb2_1229x500.png 848w, https://substackcdn.com/image/fetch/$s_!B6fz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fceb4e28d-424d-450d-a29e-598d883d1fb2_1229x500.png 1272w, https://substackcdn.com/image/fetch/$s_!B6fz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fceb4e28d-424d-450d-a29e-598d883d1fb2_1229x500.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!B6fz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fceb4e28d-424d-450d-a29e-598d883d1fb2_1229x500.png" width="1229" height="500" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ceb4e28d-424d-450d-a29e-598d883d1fb2_1229x500.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:500,&quot;width&quot;:1229,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:201003,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!B6fz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fceb4e28d-424d-450d-a29e-598d883d1fb2_1229x500.png 424w, https://substackcdn.com/image/fetch/$s_!B6fz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fceb4e28d-424d-450d-a29e-598d883d1fb2_1229x500.png 848w, https://substackcdn.com/image/fetch/$s_!B6fz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fceb4e28d-424d-450d-a29e-598d883d1fb2_1229x500.png 1272w, https://substackcdn.com/image/fetch/$s_!B6fz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fceb4e28d-424d-450d-a29e-598d883d1fb2_1229x500.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Instructions, inputs, and outputs for three tasks in Flan 2021.</figcaption></figure></div></li><li><p><strong>Flan 2022 (<a href="https://arxiv.org/abs/2210.11416">Chung et al., 2022</a><a href="https://arxiv.org/abs/2301.13688">)</a></strong>: A combination of Flan 2021, P3, Super-Natural Instructions, and additional reasoning, dialog, and program synthesis datasets. The 9 new reasoning datasets are additionally annotated with chain-of-thought (CoT; <a href="https://openreview.net/forum?id=_VjQlMeSB_J">Wei et al., 2022</a>) annotations.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IGHU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9dc9ff-f71e-41c8-8d2a-198613df6ec0_1474x602.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IGHU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9dc9ff-f71e-41c8-8d2a-198613df6ec0_1474x602.png 424w, https://substackcdn.com/image/fetch/$s_!IGHU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9dc9ff-f71e-41c8-8d2a-198613df6ec0_1474x602.png 848w, https://substackcdn.com/image/fetch/$s_!IGHU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9dc9ff-f71e-41c8-8d2a-198613df6ec0_1474x602.png 1272w, https://substackcdn.com/image/fetch/$s_!IGHU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9dc9ff-f71e-41c8-8d2a-198613df6ec0_1474x602.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IGHU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9dc9ff-f71e-41c8-8d2a-198613df6ec0_1474x602.png" width="1456" height="595" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3c9dc9ff-f71e-41c8-8d2a-198613df6ec0_1474x602.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:595,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:248273,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IGHU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9dc9ff-f71e-41c8-8d2a-198613df6ec0_1474x602.png 424w, https://substackcdn.com/image/fetch/$s_!IGHU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9dc9ff-f71e-41c8-8d2a-198613df6ec0_1474x602.png 848w, https://substackcdn.com/image/fetch/$s_!IGHU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9dc9ff-f71e-41c8-8d2a-198613df6ec0_1474x602.png 1272w, https://substackcdn.com/image/fetch/$s_!IGHU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9dc9ff-f71e-41c8-8d2a-198613df6ec0_1474x602.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Flan 2022 instructions enable fine-tuning with and without exemplars (few-shot vs zero-shot) and with and without chain-of-thought.</figcaption></figure></div></li><li><p><strong>Opt-IML Bench (<a href="https://arxiv.org/abs/2212.12017">Iyer et al., 2022</a>)</strong>: A combination of Super-Natural Instructions, P3, and Flan 2021. They additionally include dataset collections on cross-task transfer, knowledge grounding, dialogue, and a larger number of chain-of-thought datasets.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Yx0G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456a516-cb35-421f-8423-bebe296da46d_1399x597.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Yx0G!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456a516-cb35-421f-8423-bebe296da46d_1399x597.png 424w, https://substackcdn.com/image/fetch/$s_!Yx0G!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456a516-cb35-421f-8423-bebe296da46d_1399x597.png 848w, https://substackcdn.com/image/fetch/$s_!Yx0G!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456a516-cb35-421f-8423-bebe296da46d_1399x597.png 1272w, https://substackcdn.com/image/fetch/$s_!Yx0G!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456a516-cb35-421f-8423-bebe296da46d_1399x597.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Yx0G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456a516-cb35-421f-8423-bebe296da46d_1399x597.png" width="1399" height="597" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3456a516-cb35-421f-8423-bebe296da46d_1399x597.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:597,&quot;width&quot;:1399,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:187553,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Yx0G!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456a516-cb35-421f-8423-bebe296da46d_1399x597.png 424w, https://substackcdn.com/image/fetch/$s_!Yx0G!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456a516-cb35-421f-8423-bebe296da46d_1399x597.png 848w, https://substackcdn.com/image/fetch/$s_!Yx0G!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456a516-cb35-421f-8423-bebe296da46d_1399x597.png 1272w, https://substackcdn.com/image/fetch/$s_!Yx0G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3456a516-cb35-421f-8423-bebe296da46d_1399x597.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Different prompt formulations of the COPA task in Opt-IML Bench.</figcaption></figure></div></li></ol><p><a href="https://arxiv.org/abs/2301.13688">Longpre et al. (2023)</a> provide a nice overview of the timeline of some of the above datasets as well as some of their core attributes:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!v2mr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F524c44f6-f75e-4f74-a66c-bc80f1318d8a_1322x803.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!v2mr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F524c44f6-f75e-4f74-a66c-bc80f1318d8a_1322x803.png 424w, https://substackcdn.com/image/fetch/$s_!v2mr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F524c44f6-f75e-4f74-a66c-bc80f1318d8a_1322x803.png 848w, https://substackcdn.com/image/fetch/$s_!v2mr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F524c44f6-f75e-4f74-a66c-bc80f1318d8a_1322x803.png 1272w, https://substackcdn.com/image/fetch/$s_!v2mr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F524c44f6-f75e-4f74-a66c-bc80f1318d8a_1322x803.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!v2mr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F524c44f6-f75e-4f74-a66c-bc80f1318d8a_1322x803.png" width="1322" height="803" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/524c44f6-f75e-4f74-a66c-bc80f1318d8a_1322x803.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:803,&quot;width&quot;:1322,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:277918,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!v2mr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F524c44f6-f75e-4f74-a66c-bc80f1318d8a_1322x803.png 424w, https://substackcdn.com/image/fetch/$s_!v2mr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F524c44f6-f75e-4f74-a66c-bc80f1318d8a_1322x803.png 848w, https://substackcdn.com/image/fetch/$s_!v2mr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F524c44f6-f75e-4f74-a66c-bc80f1318d8a_1322x803.png 1272w, https://substackcdn.com/image/fetch/$s_!v2mr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F524c44f6-f75e-4f74-a66c-bc80f1318d8a_1322x803.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Timeline of public instruction tuning datasets (<a href="https://arxiv.org/abs/2301.13688">Longpre et al., 2023</a>).</figcaption></figure></div><div><hr></div><h3>Important Aspects of Instruction Data</h3><p><a href="https://arxiv.org/abs/2301.13688">Longpre et al. (2023)</a> and <a href="https://arxiv.org/abs/2212.12017">Iyer et al. (2022)</a> ablate several important aspects of instruction data, which we highlight in the following.</p><p><strong>Mixing few-shot settings.</strong>  Training with mixed zero-shot and few-shot prompts significantly improves performance in both settings.  </p><p><strong>Task diversity.</strong>  Large models benefit from continuously increasing the number of tasks.</p><p><strong>Data augmentation.</strong>  Augmenting the data such as by inverting inputs and outputs (e.g., turning a question answering task into a question generation task) is beneficial.</p><p><strong>Mixing weights.</strong>  When using a combination of instruction tuning datasets, appropriately tuning the mixing weights is important.</p><p>While the above datasets are mainly derived from classical NLP tasks, recent datasets such as Baize (<a href="https://arxiv.org/abs/2304.01196">Xu et al., 2023</a>), OpenAssistant Conversations (<a href="https://arxiv.org/abs/2304.07327">K&#246;pf et al., 2023</a>) and others cover a more diverse set of applications and domains. We will discuss these in the next edition. Stay tuned! &#128075;</p>]]></content:encoded></item><item><title><![CDATA[🛠 Tool-Augmented LLMs]]></title><description><![CDATA[In this edition, we&#8217;ll explore tool use&#8212;arguably one of the hottest new capabilities of LLMs. We&#8217;ll look at types of tools, benefits of tool use, recent developments, and future directions.]]></description><link>https://newsletter.ruder.io/p/tool-augmented-llms</link><guid isPermaLink="false">https://newsletter.ruder.io/p/tool-augmented-llms</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Mon, 28 Aug 2023 09:51:45 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi all,</p><p>It&#8217;s great to get back to writing regularly. Writing a newsletter provides me with an opportunity to delve into topics I&#8217;m excited about. In this edition, we&#8217;ll explore <strong>tool use</strong>&#8212;arguably one of the hottest new capabilities of LLMs. We&#8217;ll look at types of tools, benefits of tool use, recent developments, and future directions.</p><div><hr></div><p><strong>Update 24.09.23</strong>: &#127472;&#127479; This article has been <a href="https://jiho-ml.com/llm-tools-ruder/">translated to Korean by Park Ji Ho</a>. Thanks!</p><div><hr></div><h3>What is Tool Use?</h3><p>Language models are useful for a wide range of applications such as creative content generation, virtual assistants, customer support, search, etc. By definition, however, they are limited to producing natural language, which does not allow them to interact with the real world.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>This can be ameliorated by allowing the model to access <em>external tools</em>&#8212;by predicting special tokens or commands. A tool can take various forms: it can be <strong>a)</strong> the model itself or another neural network; <strong>b)</strong> a retrieval component such as a search engine; <strong>c)</strong> a symbolic computation or code module; or <strong>d)</strong> a module for controlling a physical robot or virtual agent as in the previous newsletter:</p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;bfbfbb56-55f3-40fc-9ba5-4ee7dab0cc80&quot;,&quot;caption&quot;:&quot;Hi all, This newsletter discusses components and implications for building generative agents and publication norms and venues for large language models (LLMs). NLP News is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;&#129302;&#128483; Generative Agents, &#127963; Forums for Foundation Models&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:7965403,&quot;name&quot;:&quot;Sebastian Ruder&quot;,&quot;bio&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/17fdd4c3-a575-4fe4-b58e-d876b78bfe2f_2416x2416.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2023-08-21T08:44:12.639Z&quot;,&quot;cover_image&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://nlpnewsletter.substack.com/p/generative-agents-forums-for-foundation&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:135977541,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:23,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;NLP News&quot;,&quot;publication_logo_url&quot;:&quot;https://s3.amazonaws.com/revue/profiles/images/000/027/670/thumb/sebastian_ruder_profile_photo_square.jpg?1616408613&quot;,&quot;belowTheFold&quot;:false,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><p>More broadly, a tool can be an arbitrary API. Below, are three examples of tools that can be useful for language modeling: <code>question answering</code>, <code>machine translation</code>, and a <code>calculator</code>. <a href="https://arxiv.org/abs/2302.07842">Mialon et al. (2023)</a> provide a great overview of this emerging topic in their survey.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gyBr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23afbb81-9bca-44a8-b642-4b06eb61f2a9_834x204.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gyBr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23afbb81-9bca-44a8-b642-4b06eb61f2a9_834x204.png 424w, https://substackcdn.com/image/fetch/$s_!gyBr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23afbb81-9bca-44a8-b642-4b06eb61f2a9_834x204.png 848w, https://substackcdn.com/image/fetch/$s_!gyBr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23afbb81-9bca-44a8-b642-4b06eb61f2a9_834x204.png 1272w, https://substackcdn.com/image/fetch/$s_!gyBr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23afbb81-9bca-44a8-b642-4b06eb61f2a9_834x204.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gyBr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23afbb81-9bca-44a8-b642-4b06eb61f2a9_834x204.png" width="834" height="204" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/23afbb81-9bca-44a8-b642-4b06eb61f2a9_834x204.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:204,&quot;width&quot;:834,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:77058,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gyBr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23afbb81-9bca-44a8-b642-4b06eb61f2a9_834x204.png 424w, https://substackcdn.com/image/fetch/$s_!gyBr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23afbb81-9bca-44a8-b642-4b06eb61f2a9_834x204.png 848w, https://substackcdn.com/image/fetch/$s_!gyBr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23afbb81-9bca-44a8-b642-4b06eb61f2a9_834x204.png 1272w, https://substackcdn.com/image/fetch/$s_!gyBr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23afbb81-9bca-44a8-b642-4b06eb61f2a9_834x204.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2E83!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d235f79-dd2d-43d2-944d-139eb060796e_855x129.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2E83!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d235f79-dd2d-43d2-944d-139eb060796e_855x129.png 424w, https://substackcdn.com/image/fetch/$s_!2E83!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d235f79-dd2d-43d2-944d-139eb060796e_855x129.png 848w, https://substackcdn.com/image/fetch/$s_!2E83!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d235f79-dd2d-43d2-944d-139eb060796e_855x129.png 1272w, https://substackcdn.com/image/fetch/$s_!2E83!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d235f79-dd2d-43d2-944d-139eb060796e_855x129.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2E83!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d235f79-dd2d-43d2-944d-139eb060796e_855x129.png" width="855" height="129" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3d235f79-dd2d-43d2-944d-139eb060796e_855x129.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:129,&quot;width&quot;:855,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:35275,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2E83!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d235f79-dd2d-43d2-944d-139eb060796e_855x129.png 424w, https://substackcdn.com/image/fetch/$s_!2E83!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d235f79-dd2d-43d2-944d-139eb060796e_855x129.png 848w, https://substackcdn.com/image/fetch/$s_!2E83!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d235f79-dd2d-43d2-944d-139eb060796e_855x129.png 1272w, https://substackcdn.com/image/fetch/$s_!2E83!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d235f79-dd2d-43d2-944d-139eb060796e_855x129.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HIOf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0827a03b-b6e9-48d9-95f6-20c5ea77b1fa_834x133.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HIOf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0827a03b-b6e9-48d9-95f6-20c5ea77b1fa_834x133.png 424w, https://substackcdn.com/image/fetch/$s_!HIOf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0827a03b-b6e9-48d9-95f6-20c5ea77b1fa_834x133.png 848w, https://substackcdn.com/image/fetch/$s_!HIOf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0827a03b-b6e9-48d9-95f6-20c5ea77b1fa_834x133.png 1272w, https://substackcdn.com/image/fetch/$s_!HIOf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0827a03b-b6e9-48d9-95f6-20c5ea77b1fa_834x133.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HIOf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0827a03b-b6e9-48d9-95f6-20c5ea77b1fa_834x133.png" width="834" height="133" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0827a03b-b6e9-48d9-95f6-20c5ea77b1fa_834x133.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:133,&quot;width&quot;:834,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:40807,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HIOf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0827a03b-b6e9-48d9-95f6-20c5ea77b1fa_834x133.png 424w, https://substackcdn.com/image/fetch/$s_!HIOf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0827a03b-b6e9-48d9-95f6-20c5ea77b1fa_834x133.png 848w, https://substackcdn.com/image/fetch/$s_!HIOf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0827a03b-b6e9-48d9-95f6-20c5ea77b1fa_834x133.png 1272w, https://substackcdn.com/image/fetch/$s_!HIOf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0827a03b-b6e9-48d9-95f6-20c5ea77b1fa_834x133.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Examples of different tools used by Toolformer (<a href="https://arxiv.org/abs/2302.04761">Schick et al., 2023</a>). From top to bottom: <code>question answering</code>, <code>machine translation</code>, <code>calculator</code>.</figcaption></figure></div><h3>Benefits of Tools</h3><p>Tools provide a practical way to address some of the limitations of current LLMs:</p><p>&#10060; LLMs are bad at math (e.g., <a href="https://arxiv.org/abs/2103.03874">Hendrycks et al., 2021</a>).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> &#9989; Calling a <code>calculator</code> may improve models&#8217; arithmetic capabilities.</p><p>&#10060; LLMs&#8217; pre-training data quickly becomes outdated. &#9989; Calling a <code>search engine</code> allows the LLM to produce up-to-date information.</p><p>&#10060; LLMs may hallucinate information. &#9989; Allowing an LLM to <code>cite</code> its sources may improve its trustworthiness.</p><p>&#10060; LLMs are black boxes. &#9989; A trace of the API calls an LLM used to obtain a prediction provides some degree of interpretability.</p><h3>How to Teach Tool Use</h3><p>Many tools are just an API call away&#8212;but how do we teach an LLM to use them? <strong>Few-shot prompting</strong> is a standard way to condition current models. However, a few-shot prompt may not provide enough supervision to enable an LLM to effectively use a tool, particularly if tools have complex arguments or multiple tools are required.</p><p>Instead of showing a few demonstrations of tool use to a model, we can instead provide it with <strong>tool documentation</strong>. While a demonstration showcases how a tool should be used for a specific task, documentation describes the general functionality of different tools. <a href="https://arxiv.org/abs/2308.00675">Hsieh et al. (2023)</a> find that tool documentation outperforms few-shot prompting with demonstrations on new domains.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!D_DQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc30527eb-4409-4a27-a8d9-f24909222573_1499x642.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!D_DQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc30527eb-4409-4a27-a8d9-f24909222573_1499x642.png 424w, https://substackcdn.com/image/fetch/$s_!D_DQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc30527eb-4409-4a27-a8d9-f24909222573_1499x642.png 848w, https://substackcdn.com/image/fetch/$s_!D_DQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc30527eb-4409-4a27-a8d9-f24909222573_1499x642.png 1272w, https://substackcdn.com/image/fetch/$s_!D_DQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc30527eb-4409-4a27-a8d9-f24909222573_1499x642.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!D_DQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc30527eb-4409-4a27-a8d9-f24909222573_1499x642.png" width="1456" height="624" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c30527eb-4409-4a27-a8d9-f24909222573_1499x642.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:624,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:319185,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!D_DQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc30527eb-4409-4a27-a8d9-f24909222573_1499x642.png 424w, https://substackcdn.com/image/fetch/$s_!D_DQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc30527eb-4409-4a27-a8d9-f24909222573_1499x642.png 848w, https://substackcdn.com/image/fetch/$s_!D_DQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc30527eb-4409-4a27-a8d9-f24909222573_1499x642.png 1272w, https://substackcdn.com/image/fetch/$s_!D_DQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc30527eb-4409-4a27-a8d9-f24909222573_1499x642.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Two types of information for prompting LLMs for tool use. Few-shot prompting with demonstrations (left), i.e., &lt;input, output&gt; pairs consisting of questions and their corresponding output tool-use plan. Documentation (right) provides descriptions of tool functionality (Hsieh et al., 2023). </figcaption></figure></div><p><strong>Fine-tuning on data augmented with API calls</strong> seems like the preferred choice.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> As many API calls are possible for an example, the data can be filtered to only retain &#8216;correct&#8217; API calls. In practice, an LLM can be prompted in a few-shot manner and API calls that do <em>not</em> lead to the correct final prediction are discarded. <a href="https://arxiv.org/abs/2205.12255">Parisi et al. (2022)</a>, for instance, generate sample API calls for <a href="https://research.google/pubs/pub47761/">Natural Questions</a> examples. The API calls are executed and used to produce a model response. Examples where the model produced an incorrect output are filtered and the model is fine-tuned on the updated dataset augmented with API calls.</p><p><a href="https://arxiv.org/abs/2302.04761">Schick et al. (2023)</a> use a similar strategy applied to an unlabeled text dataset (a subset of Common Crawl). Rather than only retaining API calls that lead to correct responses, they retain calls that <em>reduce the LLM&#8217;s loss</em> over the next tokens. As annotating large unlabeled texts with calls from multiple APIs is expensive, they use heuristics that inform when each API should be selected.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> </p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zQ4O!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92c5b688-7290-42e8-997b-abb4d3223603_1839x381.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zQ4O!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92c5b688-7290-42e8-997b-abb4d3223603_1839x381.png 424w, https://substackcdn.com/image/fetch/$s_!zQ4O!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92c5b688-7290-42e8-997b-abb4d3223603_1839x381.png 848w, https://substackcdn.com/image/fetch/$s_!zQ4O!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92c5b688-7290-42e8-997b-abb4d3223603_1839x381.png 1272w, https://substackcdn.com/image/fetch/$s_!zQ4O!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92c5b688-7290-42e8-997b-abb4d3223603_1839x381.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zQ4O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92c5b688-7290-42e8-997b-abb4d3223603_1839x381.png" width="1456" height="302" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/92c5b688-7290-42e8-997b-abb4d3223603_1839x381.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:302,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:195198,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zQ4O!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92c5b688-7290-42e8-997b-abb4d3223603_1839x381.png 424w, https://substackcdn.com/image/fetch/$s_!zQ4O!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92c5b688-7290-42e8-997b-abb4d3223603_1839x381.png 848w, https://substackcdn.com/image/fetch/$s_!zQ4O!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92c5b688-7290-42e8-997b-abb4d3223603_1839x381.png 1272w, https://substackcdn.com/image/fetch/$s_!zQ4O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92c5b688-7290-42e8-997b-abb4d3223603_1839x381.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Schick et al. (2023) augment an unlabeled text dataset with API calls by 1) sampling API calls for random positions in the text via few-shot prompting; 2) executing the API calls; 3) filtering out all API calls that do not reduce the LLM&#8217;s loss over the next tokens; and adding all remaining API calls to the text.</figcaption></figure></div><p>Models can also be trained using reinforcement learning with hard-coded reward functions or from human feedback (RLHF) although this may lead to instability issues during training.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><h3>Platforms for Tool-Augmented LLMs</h3><p>Given the versatility of current models, tool-augmented LLMs have quickly captured researchers&#8217; attention, with multiple recent papers claiming that tool use paves the way towards artificial general intelligence (AGI; <a href="https://arxiv.org/abs/2304.08244">Li et al., 2023</a>; <a href="https://arxiv.org/abs/2304.04370">Ge et al., 2023</a>). A central challenge for tool-augmented LLMs is the accessibility of APIs and models. The following platforms for tool-augmented LLMs have been proposed recently:</p><ul><li><p><strong><a href="http://taskmatrix.ai/">TaskMatrix.AI</a></strong> (<a href="https://arxiv.org/abs/2303.16434">March 2023</a>), a vision for an ecosystem that enables LLMs to seamlessly interface with millions of APIs. Their framework includes a base LLM, an API platform, and an API search engine. The authors envision that models mainly learn how to use APIs using RLHF, which may be difficult to scale to millions of APIs. They include a case study using ChatGPT to interface with the PowerPoint API.</p></li></ul><ul><li><p><strong>API-Bank</strong> (<a href="https://arxiv.org/abs/2304.08244">April 2023</a>), a benchmark to evaluate the tool use of LLMs in a few-shot prompting setting. In order to make tool use in the few-shot setting feasible, the model needs to produce a query for an API search engine, which returns documentation for the most relevant API.</p></li><li><p><strong><a href="https://github.com/agiresearch/OpenAGI">OpenAGI</a></strong> (<a href="https://arxiv.org/abs/2304.04370">April 2023</a>), a benchmark consisting of synthetic multi-step multi-modal datasets that require chaining calls to different domain-specific models. Models can be evaluated in zero-shot, few-shot, fine-tuning, or RL-based settings.</p></li><li><p><strong><a href="https://github.com/Gentopia-AI/Gentopia">Gentopia</a></strong> (<a href="https://arxiv.org/abs/2308.04030">August 2023</a>), a platform for creating and sharing tool-augmented agents.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LEz-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LEz-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png 424w, https://substackcdn.com/image/fetch/$s_!LEz-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png 848w, https://substackcdn.com/image/fetch/$s_!LEz-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png 1272w, https://substackcdn.com/image/fetch/$s_!LEz-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LEz-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png" width="1456" height="806" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:806,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:535136,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LEz-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png 424w, https://substackcdn.com/image/fetch/$s_!LEz-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png 848w, https://substackcdn.com/image/fetch/$s_!LEz-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png 1272w, https://substackcdn.com/image/fetch/$s_!LEz-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F695d0fa9-1845-4db0-ad45-81720e2a657e_1587x879.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Overview of TaskMatrix.AI. (1) A foundation model generates an outline of the solution based on which the API Selector selects the most relevant API. (2) The LLM generates an API call, which is executed against the API. </figcaption></figure></div><h3>Looking Back</h3><p>It is inspiring to look back to see how far the field has progressed in just a few years. There are a few trends and developments in particular that have brought us to where we are.</p><p><strong>Tool use then and now.</strong> The idea of having a model interface with auxiliary modules is not new. For instance, the Neural Programmer-Interpreter (<a href="https://arxiv.org/abs/1511.06279">Reed &amp; de Freitas, 2016</a>) required a complex neural network architecture to learn to execute different domain-specific programs; for equipping BERT with a calculator (<a href="https://arxiv.org/abs/1909.00109">Andor et al., 2019</a>), vector-based operations for a limited set of arithmetic operations were defined. What has changed is that current LLMs are much more versatile than prior models, which allows the use of arbitrary APIs.</p><p><strong>Embeddings&#8594;modules&#8594;tools.</strong> 5 years ago, we had approaches that learned to select the best combination of <em>embeddings</em> for a given task (e.g., <a href="https://aclanthology.org/D18-1176/">Kiela et al., 2018</a>). Last year, approaches selected new parameter-efficient <em>modules</em> for a given task (e.g., <a href="https://aclanthology.org/2022.acl-long.433/">Mao et al., 2022</a>). Now we are at a stage where models learn to select and use entire models and arbitrary <em>black-box tools</em>.</p><p><strong>Chitchat&#8594;goal-oriented dialogue.</strong> End-to-end goal or task-oriented dialogue has been a challenging task in NLP for a long time (<a href="https://arxiv.org/abs/1605.07683">Bordes et al., 2016</a>). While prior models have already queried database information based on their belief states (<a href="https://proceedings.neurips.cc/paper/2020/hash/e946209592563be0f01c844ab2170f0c-Abstract.html">Hosseini-Asl et al., 2020</a>), tool-augmented LLMs will be able to more seamlessly transition from chitchat to goal-oriented dialogue.</p><h3>The Future of Tool-Augmented LLMs</h3><p>Looking ahead, there are several challenges and directions for tool-augmented LLMs:</p><ol><li><p><strong>Making APIs accessible for model use.</strong> There are millions of APIs available that models can interact with. API platforms (see above) as well as <a href="https://openai.com/blog/chatgpt-plugins">ChatGPT Plugins</a> and others aim to centralize access to APIs, which may risk locking in users. To ensure research progress in this area, it will be key to ensure that a standard set of APIs are available openly and freely to use.</p></li><li><p><strong>API search and extensibility.</strong> The problem of finding the most relevant API is similar to finding the most appropriate skill for virtual assistants such as Alexa (<a href="https://aclanthology.org/P18-1206/">Kim et al., 2018</a>). It will be key to have a search component that reliably returns the most relevant API from a growing API pool as well as enabling LLMs to be easily extended with new tools.</p></li><li><p><strong>Learning to use tools.</strong> How to best teach an LLM to use tools remains an open problem. The approach of <a href="https://arxiv.org/abs/2302.04761">Schick et al. (2023)</a> is restricted to using a single tool at a time and requires tool-specific heuristics in order to augment a dataset efficiently. It will be important to investigate methods that can provide (multi-step) supervision and scale to 100s and 1000s of APIs.</p></li><li><p><strong>Pre-training tool-augmented LLMs.</strong> Given the diversity of APIs and their use cases, it makes sense to dedicate larger budgets to training tool-augmented LLMs. While pre-trained models can be fine-tuned for tool use, pre-training a tool-augmented LLM allows the model to off-load certain behavior early in training and focus on learning what is not captured by the tools. </p></li><li><p><strong>Improving reasoning and problem decomposition.</strong> Reasoning and tool use are closely intertwined (<a href="https://arxiv.org/abs/2302.07842">Mialon et al., 2023</a>). In order to call the right APIs for a problem, it needs to be decomposed into potentially simpler subtasks. How to best decompose open-ended problems is an open challenge.</p></li><li><p><strong>Compensating for API errors and preventing error cascades.</strong> API calls to other models or tools such as search engines may produce erroneous results, which can lead to downstream failures. LLMs should learn to assess the reliability of APIs and recover from API failures.</p></li><li><p><strong>Gaining a better understanding of tool use.</strong> Many aspects of how models learn to use and interface with tools are poorly understood. For instance, it is unclear to what extent models use predicted reasoning steps to support the final prediction (<a href="https://arxiv.org/abs/2212.08286">Yu et al., 2022</a>). It is thus important to develop analysis methods and diagnostic tools together with new tool-augmented LLMs.</p></li></ol><p>Overall, tool use allows us to address some of the current models&#8217; limitations and has the potential to make them more capable <em>and</em> more interpretable at the same time. I&#8217;m excited to see what future progress in this area will look like. </p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The lack of grounding to the real world, which tool use enables, has been highlighted as a limitation of LLMs in the past (<a href="https://aclanthology.org/2020.acl-main.463/">Bender &amp; Kohler, 2020</a>).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>While recent models have much improved mathematical capabilities, they are not yet able to solve graduate-level math problems (<a href="https://arxiv.org/abs/2301.13867">Frieder et al., 2023</a>). They are, however, useful as assistants to mathematicians and to guide human intuition (<a href="https://www.nature.com/articles/s41586-021-04086-x">Davies et al., 2021</a>).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>We can also refer to this as &#8216;<a href="https://www.ruder.io/recent-advances-lm-fine-tuning/#behavioural-fine-tuning">behavioral fine-tuning</a>&#8217; as we aim to teach the model something about its intended target behavior.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>For instance, texts should only be considered for the <code>calculator</code> tool if they contain at least three numbers.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>See <a href="https://arxiv.org/abs/2302.07842">Mialon et al. (2023)</a> for an overview of this area.</p></div></div>]]></content:encoded></item><item><title><![CDATA[🤖🗣 Generative Agents, 🏛 Forums for Foundation Models]]></title><description><![CDATA[This newsletter discusses components for building generative agents and publication venues for large language models (LLMs).]]></description><link>https://newsletter.ruder.io/p/generative-agents-forums-for-foundation</link><guid isPermaLink="false">https://newsletter.ruder.io/p/generative-agents-forums-for-foundation</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Mon, 21 Aug 2023 08:44:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TtNQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi all,</p><p>This newsletter discusses components and implications for building generative agents and publication norms and venues for large language models (LLMs). </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">NLP News is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>&#129302;&#128483; Generative Agents</h2><p>Throughout the history of AI, people have been fascinated with the prospect of interacting with AI agents. Traditionally, interactions have been restricted to short time horizons and individual conversations such as with <a href="https://en.wikipedia.org/wiki/ELIZA">ELIZA</a> or the <a href="https://en.wikipedia.org/wiki/Turing_test">Turing test</a>. The rise of large language models (LLMs) has made it easier to develop persona-based bots (with <a href="https://www.howtogeek.com/881659/how-to-create-chatgpt-personas-for-every-occasion/">ChatGPT</a> or <a href="https://beta.character.ai/">Character.ai</a>, for instance) that have a consistent personality but these bots similarly adhere to the persona one conversation at a time.</p><p>At the same time, we have already seen glimpses of LLM-powered world-building in applications such as <a href="https://aidungeon.io/">AI dungeon</a>. While it is possible to interact with non-player characters and events in text adventures, characters only exist in relation to the player. <strong>What does it take then to generate a world where agents pursue their own goals and objectives regardless of player input?</strong> It turns out, LLMs (+ a game engine) are all you need&#8212;at least for the small town sandbox created by Park et al. in <a href="https://arxiv.org/abs/2304.03442">Generative Agents: Interactive Simulacra of Human Behavior</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TtNQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TtNQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png 424w, https://substackcdn.com/image/fetch/$s_!TtNQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png 848w, https://substackcdn.com/image/fetch/$s_!TtNQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png 1272w, https://substackcdn.com/image/fetch/$s_!TtNQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TtNQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png" width="1456" height="779" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:779,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1364421,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TtNQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png 424w, https://substackcdn.com/image/fetch/$s_!TtNQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png 848w, https://substackcdn.com/image/fetch/$s_!TtNQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png 1272w, https://substackcdn.com/image/fetch/$s_!TtNQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f2c7c4c-72eb-48b2-8245-99c018d858c3_1524x815.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The Smallville town sandbox where LLM agents go about their day and interact with each other. An LLM is used to convert an action description to an emoji sequence that is displayed to the user (Park et al., 2023).</figcaption></figure></div><p>In the paper, Park et al. simulate a population of 25 LLM agents in a The Sims-like sandbox environment. Each agent consists of a one-paragraph natural language description depicting its identity and multiple components, each enabled by a LLM (ChatGPT)<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>:</p><p><strong>Memory and retrieval</strong>. A memory stream records events perceived by the agent (with a natural language description, event timestamp, and most recent access timestamp). A retriever retrieves a set of memory objects using a query description based on three factors:</p><ol><li><p><em>recency</em> (based on exponential decay since the memory was last retrieved);</p></li><li><p><em>importance</em> (the LLM assigns an importance score based on the memory&#8217;s description);</p></li><li><p>and <em>relevance</em> (based on embedding similarity with the query).</p></li></ol><p>It is nice to see the return of explicit memory mechanisms, similar in spirit to the key-value memory of <a href="https://arxiv.org/abs/1410.5401">Neural Turing Machines</a> and later work. An advantage of this type of memory is interpretability: The contents of the memory of each agent are human-readable and can be inspected at every point in time.</p><p><strong>Reflection.</strong> Beyond simply remembering past events, an agent should be able to reflects on its experiences and generate higher-level thoughts. To this end, the LLM is queried using the 100 most recent memory records and asked to suggest 3 high-level questions to ask about the subjects in the memory records. For each question, memory records are then retrieved and the model is asked to generate high-level insights based on the retrieved records. The generated insights are then added to the memory, enabling the agent to recursively generate higher-level observations about prior reflections.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R-4x!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19dd1402-06c1-4393-bc3c-204a3be0b773_1539x772.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R-4x!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19dd1402-06c1-4393-bc3c-204a3be0b773_1539x772.png 424w, https://substackcdn.com/image/fetch/$s_!R-4x!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19dd1402-06c1-4393-bc3c-204a3be0b773_1539x772.png 848w, https://substackcdn.com/image/fetch/$s_!R-4x!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19dd1402-06c1-4393-bc3c-204a3be0b773_1539x772.png 1272w, https://substackcdn.com/image/fetch/$s_!R-4x!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19dd1402-06c1-4393-bc3c-204a3be0b773_1539x772.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R-4x!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19dd1402-06c1-4393-bc3c-204a3be0b773_1539x772.png" width="1456" height="730" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/19dd1402-06c1-4393-bc3c-204a3be0b773_1539x772.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:730,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:372462,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!R-4x!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19dd1402-06c1-4393-bc3c-204a3be0b773_1539x772.png 424w, https://substackcdn.com/image/fetch/$s_!R-4x!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19dd1402-06c1-4393-bc3c-204a3be0b773_1539x772.png 848w, https://substackcdn.com/image/fetch/$s_!R-4x!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19dd1402-06c1-4393-bc3c-204a3be0b773_1539x772.png 1272w, https://substackcdn.com/image/fetch/$s_!R-4x!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19dd1402-06c1-4393-bc3c-204a3be0b773_1539x772.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">A reflection tree for Klaus Mueller. Reflections such as &#8220;Klaus Mueller is dedicated to research&#8221; are generated based on observations from the memory and can be generated recursively based on prior reflections.</figcaption></figure></div><p><strong>Planning and reacting.</strong> To plan what an agent does each day, the authors generate a plan for each day recursively by first generating a rough sketch based on the agent&#8217;s description and a summary of their previous day. The LLM decomposes the plan then first into hour-long and again into 5&#8211;15 minute chunks.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> During the day, the LLM is prompted with the perceived observations to decide if and how it should react to a situation. If a reaction is to engage in dialogue with another agent, then the LLMs generate the utterances conditioned on their memories about each other and the previous dialogue history.</p><p>LLMs are truly ubiquitous in this work: They are used for retrieval, to assign importance to memories, to suggest what to reflect on, to generate insights, to convert actions to emojis, to generate plans and dialogue, etc. Even the game environment itself is powered by LLMs: In order to determine the appropriate location for each action, the LLM is queried recursively. When an agent executes an action on an object, the LLM is asked about the state of the object. If anything, this demonstrates the versatility of current models.</p><p>It is sobering that as a field, we are at a point where many elements of a complex ML agent stack can be replaced with an LLM and an appropriately worded prompt. Most of the above use cases required training a specialized system not long ago, from producing descriptions and explanations of video game actions (<a href="https://arxiv.org/abs/1901.03729">Ehsan et al., 2019</a>) to mapping from text to emoji (<a href="https://aclanthology.org/W16-6208/">Eisner et al., 2016</a>). Of course, it will be useful to specialize the model to improve performance for specific use cases such as retrieval or dialogue. While the authors ablate each of the above three components via a human evaluation, I am missing a more fine-grained analysis that highlights for which use cases the LLM is most brittle, which can inform future investigations.</p><p>It is unsurprising that the project in its current form costs thousands of dollars to run just with 25 agents. With smaller open-source models, the costs should drop dramatically. The code for the framework in the paper can be found <a href="https://github.com/joonspk-research/generative_agents">here</a>. Inspired by the paper, researchers from Andreessen Horowitz also made available <a href="https://github.com/a16z-infra/AI-town">AI Town</a>, an MIT-licensed starter kit for building and customizing your own virtual town populated by AI characters.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LRRi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7edef242-a007-4eec-aad6-0086586676ea_2908x1862.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LRRi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7edef242-a007-4eec-aad6-0086586676ea_2908x1862.png 424w, https://substackcdn.com/image/fetch/$s_!LRRi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7edef242-a007-4eec-aad6-0086586676ea_2908x1862.png 848w, https://substackcdn.com/image/fetch/$s_!LRRi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7edef242-a007-4eec-aad6-0086586676ea_2908x1862.png 1272w, https://substackcdn.com/image/fetch/$s_!LRRi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7edef242-a007-4eec-aad6-0086586676ea_2908x1862.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LRRi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7edef242-a007-4eec-aad6-0086586676ea_2908x1862.png" width="1456" height="932" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7edef242-a007-4eec-aad6-0086586676ea_2908x1862.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:932,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1957247,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LRRi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7edef242-a007-4eec-aad6-0086586676ea_2908x1862.png 424w, https://substackcdn.com/image/fetch/$s_!LRRi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7edef242-a007-4eec-aad6-0086586676ea_2908x1862.png 848w, https://substackcdn.com/image/fetch/$s_!LRRi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7edef242-a007-4eec-aad6-0086586676ea_2908x1862.png 1272w, https://substackcdn.com/image/fetch/$s_!LRRi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7edef242-a007-4eec-aad6-0086586676ea_2908x1862.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">AI Town, an open-source framework for building virtual towns populated by AI agents.</figcaption></figure></div><p>The paper acts as a blueprint of how a framework of AI agents interacting may look like. It is worth pointing out that current models have a range of biases and may exhibit other forms of undesirable behavior that require further study. Nevertheless, there are compelling research directions in the area of multi-agent LLM systems:</p><ol><li><p><strong>Emergent communication.</strong> The emergence of communicative behavior between AI agents has been mainly studied using simple reference games (<a href="https://openreview.net/forum?id=AUGBfDIV9rL">Chaabouni et al., 2022</a>). What can richer simulated environments tell us about how agents learn to communicate? Can we simulate emergent communication in other settings such as across different languages?</p></li><li><p><strong>Nature of communication.</strong> Can we use these environments to study aspects of how humans communicate such as <a href="http://www.ello.uos.de/field.php/Sociolinguistics/Linguisticaccommodation">speaker accommodation</a>, <a href="https://www.sciencedirect.com/science/article/abs/pii/0378216694900132">register change</a>, and <a href="https://en.wikipedia.org/wiki/Code-switching">code-switching</a>? </p></li><li><p><strong>Nature of human behavior and social interaction.</strong> Can we study factors involved in other types of higher-level behavior such as planning, organization, co-operation, collaboration, and deception?</p></li><li><p><strong>Multi-modal environments.</strong> Can we extend environments to incorporate audio and visual inputs and assess how these additional modalities affect the agents&#8217; behavior?</p></li><li><p><strong>The expressiveness of simulations.</strong> What kind of methods are necessary to simulate even richer interactions? What model scale and capabilities would be necessary to simulate agents and events at the scale of our world?</p></li></ol><h2>&#127963; Forums for Foundation Models</h2><p>There was some <a href="https://twitter.com/denny_zhou/status/1689366382566326272">discussion on Twitter</a> recently about initiating a new conference dedicated specifically to LLMs. LLMs are an interesting technology because they are at the intersection of many different areas of computer science and society including NLP and ML, human-computer interaction, ethics, law, government, education, etc.</p><p>So far, work on LLMs has been published in a range of different venues. Much of the foundational work on LLMs (for example, <a href="https://aclanthology.org/N18-1202/">ELMo</a>, <a href="https://aclanthology.org/P18-1031/">ULMFiT</a>, <a href="https://aclanthology.org/N19-1423/">BERT</a>) has been published in NLP (*ACL) conferences, which remain the most topically relevant venue.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> The NLP community has studied topics, which are of increasing importance for LLMs, such as automatic evaluation of generative models (<a href="https://arxiv.org/abs/2202.06935">Gehrmann et al., 2022</a>) for decades. Recently, <a href="https://twitter.com/xwang_lk/status/1689709900312756225?s=20">researchers</a> have taken issue with the anonymity period<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> but there are <a href="https://docs.google.com/forms/d/e/1FAIpQLSdCYqojzzW-vtjuOryybdziF0ywPSYBJOZYRGT6WWvGwQH7CA/viewform">initiatives</a> to rethink it.</p><p>LLMs have also increasingly become more popular at ML venues. <a href="https://nips.cc/virtual/2020/oral/18545">GPT-3</a> won an outstanding paper award at NeurIPS 2020 while NeurIPS 2022 featured outstanding papers on <a href="https://nips.cc/virtual/2022/poster/53031">scaling laws</a> and <a href="https://nips.cc/virtual/2022/poster/55659">large-scale image&#8211;text datasets</a>. There are also a range of dedicated smaller venues such as the <a href="https://sites.google.com/mila.quebec/scaling-laws-workshop?pli=1">Neural Scaling Laws workshop series</a>. There is even a new conference on <a href="https://www.iaria.org/conferences2024/ComGPTMB24.html">Generative Pre-trained Transformer Models and Beyond</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><p>The above venues mainly cater to a computer science audience. As LLMs are increasingly used in user-facing applications, work on LLMs is also published in venues associated with other areas such as the <a href="https://arxiv.org/abs/2304.03442">ACM Symposium on User Interface Software and Technology</a> for <a href="https://arxiv.org/abs/2304.03442">the paper</a> in the previous section. Given the societal impact of LLMs, it is thus important for any venue to enable an inter-disciplinary dialogue that ensures LLMs are developed in a safe, responsible, and user-appropriate manner.</p><p>Besides <em>where</em> to publish, the other issue is <em>who</em> can publish. At the moment, much of the work on LLMs is conducted by labs with large compute budgets. For the previous generation of pre-trained models, an entire array of papers from a diverse range of institutions focused on gaining a better understanding of them.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> Such work is needed for the latest generation of large models but is currently prohibitive for many labs. With the release of new powerful open-source models such as <a href="https://arxiv.org/abs/2307.09288">Llama-2</a> combined with better methods for <a href="https://arxiv.org/abs/2208.07339">compute-efficient inference</a>, we are heading in a promising direction, however.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The broader family of LLMs that can learn to act and use auxiliary modules is known as <a href="https://arxiv.org/abs/2302.07842">augmented language models</a>. We will discuss these more in-depth in the next edition.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>We have also recently seen plans in other areas such as summarization where question-answer pairs have been used as intermediate representation for conditional generation (<a href="https://arxiv.org/abs/2207.00397">Narayan et al., 2022</a>), for instance.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>ACL 2023 had a <a href="https://2023.aclweb.org/calls/main_conference/">track dedicated to LLMs</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p><a href="https://www.aclweb.org/adminwiki/index.php/ACL_Policies_for_Submission,_Review_and_Citation">Submitted papers are not allowed</a> to be posted, updated, or discussed online from 1 month before the submission deadline until the time of notification (around 4 months in total). </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>However, a focus on a specific architecture (and a name associated with a specific model family) may be potentially restrictive in the future.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>This area of study has been commonly referred to as <a href="https://aclanthology.org/2020.tacl-1.54/">BERTology</a>.</p></div></div>]]></content:encoded></item><item><title><![CDATA[✨ Flashier Attention, 🤐 Gzip Classifiers]]></title><description><![CDATA[This newsletter discusses attention that enables modeling long sequences and simple but surprisingly competitive classifiers such as based on gzip compression.]]></description><link>https://newsletter.ruder.io/p/flashier-attention-gzip-classifiers</link><guid isPermaLink="false">https://newsletter.ruder.io/p/flashier-attention-gzip-classifiers</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Mon, 14 Aug 2023 08:07:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi all,</p><p>This newsletter has been on hiatus for some time due to some personal developments (I became a father recently &#128118;). So besides tackling an avalanche of dirty nappies, I&#8217;ll also wade again through the sea of arXiv research papers.</p><p>As before, in each edition, I will analzye a couple of topics (each consisting of one or multiple papers) by contextualizing them and reflecting on what they entail for the future of NLP and large language models (LLMs).</p><p><em>Note: If you enjoy synthesizing information and connecting your work to the broader research manifold, then you might also be interested in <a href="https://www.bigpictureworkshop.com/">The Big Picture Workshop</a>, which I&#8217;m co-organizing at EMNLP 2023 (submission deadline: September 1). </em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading NLP News! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>&#10024; Flashier Attention: Towards long-range models via hardware optimization</h2><p>The current generation of large language models is based on the Transformer architecture. One of the limitations of Transformers is that the time and memory complexity of self-attention, one of its core building blocks, is quadratic in the sequence length. <strong>This makes scaling to very long sequences such as books, many documents, or rich interaction data prohibitive.</strong></p><p>Numerous efficient attention methods have been proposed over the years (see <a href="https://arxiv.org/abs/2009.06732">this survey</a> for an overview). Most of these methods seek to reduce the number of operations by approximating the attention algorithm, for instance, by sparsifying the attention matrix. However, the <em>theoretical</em> efficiency gains of these methods rarely translate into practical speed-ups as sparse-matrix operations are still poorly supported by current accelerators.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!V8vg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e9cf474-c35c-4240-bbd9-a74d2ed16829_1746x1290.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!V8vg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e9cf474-c35c-4240-bbd9-a74d2ed16829_1746x1290.png 424w, https://substackcdn.com/image/fetch/$s_!V8vg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e9cf474-c35c-4240-bbd9-a74d2ed16829_1746x1290.png 848w, https://substackcdn.com/image/fetch/$s_!V8vg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e9cf474-c35c-4240-bbd9-a74d2ed16829_1746x1290.png 1272w, https://substackcdn.com/image/fetch/$s_!V8vg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e9cf474-c35c-4240-bbd9-a74d2ed16829_1746x1290.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!V8vg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e9cf474-c35c-4240-bbd9-a74d2ed16829_1746x1290.png" width="728" height="538" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8e9cf474-c35c-4240-bbd9-a74d2ed16829_1746x1290.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:1076,&quot;width&quot;:1456,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:567306,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!V8vg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e9cf474-c35c-4240-bbd9-a74d2ed16829_1746x1290.png 424w, https://substackcdn.com/image/fetch/$s_!V8vg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e9cf474-c35c-4240-bbd9-a74d2ed16829_1746x1290.png 848w, https://substackcdn.com/image/fetch/$s_!V8vg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e9cf474-c35c-4240-bbd9-a74d2ed16829_1746x1290.png 1272w, https://substackcdn.com/image/fetch/$s_!V8vg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e9cf474-c35c-4240-bbd9-a74d2ed16829_1746x1290.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">A taxonomy of efficient Transformer architectures (Tay et al., 2022). </figcaption></figure></div><p>FlashAttention (<a href="https://arxiv.org/abs/2205.14135">arXiv May &#8216;22</a>) is a method that takes into account the <em>hardware</em> constraints of attention and reduces the number of times the slow GPU memory needs to be accessed during the attention computation. Specifically, it <strong>a)</strong> computes the softmax incrementally and <strong>b)</strong> stores the softmax normalization factor for faster recomputation in the backward pass. For a more detailed explanation, check out <a href="https://gordicaleksa.medium.com/eli5-flash-attention-5c44017022ad">this article</a> by Aleksa Gordi&#263;.</p><p>FlashAttention achieves speed-ups between 2&#8211;3x for different Transformer models and has enabled scaling Transformers to longer sequence lengths. It is the first method that achieved non-random performance on the challenging Path-X task of <a href="https://github.com/google-research/long-range-arena">Long-Range Arena</a>, a benchmark for assessing efficient Transformers. It is now available as a plug-in replacement for attention in <a href="https://github.com/Dao-AILab/flash-attention/blob/main/usage.md">many ML frameworks</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gkR4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07aa5635-889a-458d-adc2-bd354408f0a9_2218x1098.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gkR4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07aa5635-889a-458d-adc2-bd354408f0a9_2218x1098.jpeg 424w, https://substackcdn.com/image/fetch/$s_!gkR4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07aa5635-889a-458d-adc2-bd354408f0a9_2218x1098.jpeg 848w, https://substackcdn.com/image/fetch/$s_!gkR4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07aa5635-889a-458d-adc2-bd354408f0a9_2218x1098.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!gkR4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07aa5635-889a-458d-adc2-bd354408f0a9_2218x1098.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gkR4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07aa5635-889a-458d-adc2-bd354408f0a9_2218x1098.jpeg" width="1456" height="721" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/07aa5635-889a-458d-adc2-bd354408f0a9_2218x1098.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:721,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:258323,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gkR4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07aa5635-889a-458d-adc2-bd354408f0a9_2218x1098.jpeg 424w, https://substackcdn.com/image/fetch/$s_!gkR4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07aa5635-889a-458d-adc2-bd354408f0a9_2218x1098.jpeg 848w, https://substackcdn.com/image/fetch/$s_!gkR4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07aa5635-889a-458d-adc2-bd354408f0a9_2218x1098.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!gkR4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07aa5635-889a-458d-adc2-bd354408f0a9_2218x1098.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">A positive and negative example of Long-Range Arena&#8217;s Pathfinder task, which requires deciding whether two points are connected by a path of dashes. Path-X scales Pathfinder to sequences of length 16k (i.e., 128x128 images), a sequence length that was until recently prohibitive for existing approaches. </figcaption></figure></div><p>Recently, Tri Dao, the first author of FlashAttention, proposed FlashAttention-2 (<a href="https://arxiv.org/abs/2307.08691">arXiv July &#8216;23</a>), which offers additional hardware optimizations by <strong>a)</strong> reducing the number of FLOPs; <strong>b)</strong> parallelizing not only over batch size and number of heads but also over sequence length; and <strong>c)</strong> further partitioning operations within each thread.</p><p>The second optimization in particular enables scaling to very long sequence lengths (as Trio Dao explains in <a href="https://www.adept.ai/blog/flashier-attention">this blog post</a>). Methods like FlashAttention-2 that make modeling of long sequences feasible have several implications for LLMs and efficient Transformers:</p><ol><li><p><strong>Models will be able to use much longer inputs. </strong>This lifts restrictions on users and developers. For instance, developers no longer need to artificially split or truncate long documents or come up with elaborate ways to reconcile model outputs from different subsets of the input.</p></li><li><p><strong>Existing applications will get revisited and updated.</strong> Many existing NLP applications such as multi-document summarization and open-domain question answering have been subject to input length restrictions. New benchmarks and methods will reframe these tasks in light of long-context modeling.</p></li><li><p><strong>Long-sequence modeling enables new applications and research directions.</strong> Modeling of extremely long sequences such as entire books, rich user interaction, and other forms of longitudinal sequential data is now possible. Credit assignment over such long sequences is still a challenge, however, as well as understanding and optimizing long-range memory (see, e.g., <a href="http://aclanthology.lst.uni-saarland.de/2020.acl-main.672/">Rae &amp; Razavi, 2020</a>). Long-range sequence modeling may also be a boon for more granular representations such as character-level models (<a href="https://aclanthology.org/2022.tacl-1.5/">Clark et al., 2022</a>; <a href="https://openreview.net/forum?id=JtBRnrlOEFN">Tay et al., 2022</a>), which have been hindered by limits on sequence length.</p></li><li><p><strong>Hardware constraints will play a bigger role in influencing future advances. </strong>This has been the case throughout the history of ML but will become more pronounced in light of the huge compute budgets necessary to train current models. We will see more methods that achieve gains in efficiency through clever hardware utilization; some recent examples: <a href="https://proceedings.neurips.cc/paper/2020/hash/81f7acabd411274fcf65ce2070ed568a-Abstract.html">TinyTL</a> (memory reduction by only updating biases), <a href="https://arxiv.org/abs/2106.09685">LoRA</a> (adapters without additional inference latency), and <a href="https://arxiv.org/abs/2305.14314">QLoRA</a> (quantized adapters).</p></li></ol><div><hr></div><h2>&#129296; Gzip: The case for simple models</h2><p>In light of the excitement around pre-trained Transformer models, it may be easy to forget that there are a huge range of other architectures and models, which may be more suitable depending on what considerations are important for a given use case.</p><p>I am generally a fan of papers that diverge from the beaten track and explore or revisit under-explored methods. During my PhD, for instance, I explored classic semi-supervised learning methods (<a href="https://aclanthology.org/P18-1096/">ACL 2018</a>). Another paper in this vein that I enjoyed is by Tay et al. (<a href="https://aclanthology.org/2021.acl-long.335/">ACL 2021</a>) who compared pre-trained convolutions to pre-trained Transformers.</p><p>Recently, another contrarian paper attracted attention on Twitter. It proposes using a gzip-based compression and distance metric with a <em>k</em>-nearest-neighbor classifier for text classification (<a href="https://aclanthology.org/2023.findings-acl.426/">Findings of ACL 2023</a>).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> Remarkably, the authors show competitive performance with more complex non-pre-trained methods on six text classification datasets and even seemingly outperform BERT on five out-of-distribution datasets.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yff9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108288a9-9ad5-4790-948d-f8b45def65d8_2164x578.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yff9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108288a9-9ad5-4790-948d-f8b45def65d8_2164x578.png 424w, https://substackcdn.com/image/fetch/$s_!yff9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108288a9-9ad5-4790-948d-f8b45def65d8_2164x578.png 848w, https://substackcdn.com/image/fetch/$s_!yff9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108288a9-9ad5-4790-948d-f8b45def65d8_2164x578.png 1272w, https://substackcdn.com/image/fetch/$s_!yff9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108288a9-9ad5-4790-948d-f8b45def65d8_2164x578.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yff9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108288a9-9ad5-4790-948d-f8b45def65d8_2164x578.png" width="1456" height="389" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/108288a9-9ad5-4790-948d-f8b45def65d8_2164x578.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:389,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:201939,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yff9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108288a9-9ad5-4790-948d-f8b45def65d8_2164x578.png 424w, https://substackcdn.com/image/fetch/$s_!yff9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108288a9-9ad5-4790-948d-f8b45def65d8_2164x578.png 848w, https://substackcdn.com/image/fetch/$s_!yff9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108288a9-9ad5-4790-948d-f8b45def65d8_2164x578.png 1272w, https://substackcdn.com/image/fetch/$s_!yff9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F108288a9-9ad5-4790-948d-f8b45def65d8_2164x578.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Test accuracy of the gzip-based method compared to other text classification methods on out-of-distribution (OOD) datasets (Jiang et al., 2023).</figcaption></figure></div><p>The authors set a good example by releasing the source code of their method, enabling others to experiment and to try to reproduce their results. However, researchers such as Ken Schutte quickly identified some inconsistencies with their method. Specifically, the <a href="https://kenschutte.com/gzip-knn-paper/">authors report an &#8216;oracle&#8217; top-2 accuracy for their method</a> rather than a standard accuracy as used by the baselines; in addition, there are <a href="https://kenschutte.com/gzip-knn-paper2/">issues with the underlying source datasets</a>, which lead to unreliable results while computing the distance metric for large training sets is excessively slow. <a href="https://twitter.com/nlopitz/status/1689267192095703040?s=20">Juri Opitz</a> also observes that a <a href="https://arxiv.org/abs/2307.15002">bag-of-words-based distance metric outperforms</a> gzip-based  distance.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>Unfortunately, as usual, there is &#8216;no free lunch&#8217; in machine learning&#8212;at least when it comes to non-pre-trained models. It is unlikely that a simple method such as gzip-based compression outperforms pre-trained models that have learned much richer representations unless it has been designed with specific use cases in mind.</p><p>Nevertheless, simplicity is a virtue. <strong>The promise of simple models is interpretability, which has proven elusive in the current era of black-box behemoths.</strong> Two of my favorite &#8216;classic&#8217; examples of deceptively simple methods that punch above their weight are the Deep Averaging Network (<a href="https://aclanthology.org/P15-1162/">ACL 2015</a>) and the feature-based classifier used to examine the CNN/Daily Mail dataset (<a href="https://aclanthology.org/P16-1223/">ACL 2016</a>). The latter demonstrates the advantage of interpretability, highlighting that simple features are sufficient for many reading comprehension questions. </p><p>A family of models that expertly embodies this notion of simplicity and interpretability are Generalized Additive Models (GAMs)<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>, which consist of a linear combination of single-feature functions:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;g(\\mathbb{E}[y]) = &#946; + f_1(x_1) + f_2(x_2) + \\ldots + f_K(x_K) &quot;,&quot;id&quot;:&quot;VPJRUEXXMT&quot;}" data-component-name="LatexBlockToDOM"></div><p>where <em>g</em> is the &#8216;link&#8217; function such as a logistic function. Recently, Neural Additive Models (NAMs; <a href="https://arxiv.org/abs/2004.13912">NeurIPS 2021</a>), a neural extension of GAMs have been proposed, which learn a linear combination of single-feature <em>neural networks</em>. We have not seen much of these models in practice for NLP applications due to the high dimensionality of language features and the complexity of natural language.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!r-pb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!r-pb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png 424w, https://substackcdn.com/image/fetch/$s_!r-pb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png 848w, https://substackcdn.com/image/fetch/$s_!r-pb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png 1272w, https://substackcdn.com/image/fetch/$s_!r-pb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!r-pb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png" width="1456" height="846" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/de66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:846,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:416168,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!r-pb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png 424w, https://substackcdn.com/image/fetch/$s_!r-pb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png 848w, https://substackcdn.com/image/fetch/$s_!r-pb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png 1272w, https://substackcdn.com/image/fetch/$s_!r-pb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde66d733-eaa7-4d76-aea0-1954664ee3c5_2126x1236.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">An example of understanding individual NAM predictions for credit scores. For an input, each feature net returns a contribution term, which are added up and passed through the link function for prediction, allowing for easy interpretability.</figcaption></figure></div><p>Nevertheless, as we continue to design models that are most useful to real-world users, it is important to remember that huge pre-trained models may not be the silver bullet for every problem. We should be aware that there are glass-box models that are more interpretable and efficient and may thus be a better fit for applications where these characteristics are important.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>How to optimally compress text is a long-standing problem. The <a href="http://prize.hutter1.net/">Hutter Prize</a>, for instance, asks participants to compress enwik9 as much as possible. <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Abhinav Upadhyay&quot;,&quot;id&quot;:14520974,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c6dafc3-e0d4-40db-9510-4c57f3bc8e13_1532x1532.jpeg&quot;,&quot;uuid&quot;:&quot;434c1212-f5e5-4b33-9fa4-f604b0358057&quot;}" data-component-name="MentionToDOM"></span> compares alternative compression algorithms to gzip in <a href="https://codeconfessions.substack.com/p/lz77-is-all-you-need">this post</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>It is important to remind ourselves that this follow-up work was mainly possible because the authors released their code&#8212;kudos to them!</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Generalized linear models such as logistic regression are a special form of GAMs.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Moving to Substack ➡️, Scaling Up 📈, Image Generation 🖼]]></title><description><![CDATA[Hi all,]]></description><link>https://newsletter.ruder.io/p/moving-to-substack-scaling-up-image</link><guid isPermaLink="false">https://newsletter.ruder.io/p/moving-to-substack-scaling-up-image</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Sun, 06 Nov 2022 21:31:50 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!KbvJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi all,</p><p>Welp. Twitter <a href="https://www.theinformation.com/briefings/twitter-will-shut-down-newsletter-product-revue-by-year-end">plans to shut down Revue</a>, my previous newsletter platform. I hope you like our new home, <a href="https://substack.com/">Substack</a>. &#129303; I&#8217;m actually quite excited about the change. Substack allows comments on posts. We can now have actual discussions rather than just one-on-one email exchanges. Yay!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading NLP News! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>If you are already subscribed, everything should stay the same (I hope). If you are not, have a look around and see if you like things. All newsletter issues will stay free as before.</p><p>Here is some coverage of recent work on <strong>scaling up language models</strong> and <strong>text-based image generation</strong> that I did not get around to publishing (updated with some recent results). </p><h1>Scaling up results &#128200;</h1><p>As language models have become larger, <a href="https://ruder.io/nlp-benchmarking/">they have outgrown some of the datasets we used to evaluate them</a>. <a href="https://arxiv.org/abs/2206.04615">BIG-Bench</a>, a two-year collaboration consisting of 204 tasks created by 442 authors aims to provide a diverse collection of tasks for the evaluation of current and future models.</p><p>Using 204 tasks for evaluation can be quite unwieldy, however, so recent work has already focused on <a href="https://arxiv.org/abs/2210.09261">23 particularly challenging tasks</a> where prior models did not outperform human annotators. It turns out that if you use <a href="https://arxiv.org/abs/2201.11903">chain-of-thought prompting</a>&#8212;a recent method where you prompt a method to predict intermediate reasoning steps before producing the final answer&#8212;then recent large models surpass the average human annotator on up to 17/23 tasks. It seems we need to keep looking for more challenging evaluation tasks for large language models.</p><p>One thing that we have learned from training and evaluating large language models, is that "scale generally works". In other words, making models larger and training on more data usually leads to improved performance.</p><p>In some cases, this can lead to new behaviour. A small model may be bad at adding two numbers while a larger model may suddenly be able to perform the task with reasonable accuracy. In such cases, the linear performance curves that we would expect based on <a href="https://arxiv.org/abs/2001.08361">scaling laws</a> are instead step-functions that transition from random to above-random performance at a certain model scale.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KbvJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KbvJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg 424w, https://substackcdn.com/image/fetch/$s_!KbvJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg 848w, https://substackcdn.com/image/fetch/$s_!KbvJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!KbvJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KbvJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg" width="740" height="330" data-attrs="{&quot;src&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/cec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:330,&quot;width&quot;:740,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:28224,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KbvJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg 424w, https://substackcdn.com/image/fetch/$s_!KbvJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg 848w, https://substackcdn.com/image/fetch/$s_!KbvJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!KbvJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fcec2e37d-a4d8-499d-a66e-44ebf1832d49_740x330.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Emergent few-shot abilities of recent large language models on three benchmarks. 3-digit addition and subtraction, 2-digit multiplication (left); transliteration from the International Phonetic Alphabet (IPA) (middle); performance on the Word in Context (WiC) task (right) (<a href="https://arxiv.org/abs/2206.07682">Wei et al., 2022</a>).</figcaption></figure></div><p>Such behaviour has been observed by recent large model papers such as <a href="https://arxiv.org/abs/2201.08239">LaMDA</a>, <a href="https://arxiv.org/abs/2005.14165">GPT-3</a>, <a href="https://arxiv.org/pdf/2112.11446.pdf">Gopher</a>, or <a href="https://arxiv.org/abs/2204.02311">PaLM</a>. <a href="https://arxiv.org/abs/2206.07682">Wei et al. (2022)</a> provide an overview of other types of emergent abilities that have been identified so far in current models.</p><p>Besides such emergent abilities, is there any other behaviour that may be surprising as we scale models up? The <a href="https://github.com/inverse-scaling/prize">Inverse Scaling Prize</a>, a recent competition awarded a prize of up to $100,000 if you find a task where performance goes down &#128201; as models get larger.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-05y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F0d404b5a-a919-41f8-a806-fe7a35c3768c_740x387.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-05y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F0d404b5a-a919-41f8-a806-fe7a35c3768c_740x387.jpeg 424w, https://substackcdn.com/image/fetch/$s_!-05y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F0d404b5a-a919-41f8-a806-fe7a35c3768c_740x387.jpeg 848w, https://substackcdn.com/image/fetch/$s_!-05y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F0d404b5a-a919-41f8-a806-fe7a35c3768c_740x387.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!-05y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F0d404b5a-a919-41f8-a806-fe7a35c3768c_740x387.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-05y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F0d404b5a-a919-41f8-a806-fe7a35c3768c_740x387.jpeg" width="740" height="387" data-attrs="{&quot;src&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/0d404b5a-a919-41f8-a806-fe7a35c3768c_740x387.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:387,&quot;width&quot;:740,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:30678,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-05y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F0d404b5a-a919-41f8-a806-fe7a35c3768c_740x387.jpeg 424w, https://substackcdn.com/image/fetch/$s_!-05y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F0d404b5a-a919-41f8-a806-fe7a35c3768c_740x387.jpeg 848w, https://substackcdn.com/image/fetch/$s_!-05y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F0d404b5a-a919-41f8-a806-fe7a35c3768c_740x387.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!-05y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F0d404b5a-a919-41f8-a806-fe7a35c3768c_740x387.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>A good way to narrow the search space is to look for tasks where an inductive bias is more important than more parameters. A minimal example for this is highlighted by <a href="https://twitter.com/hardmaru/status/1541656926647623680">@hardmaru</a>. Given that the evaluation will focus on large language models, anything that <em>cannot</em> be learned from large amounts of text may be a good starting point. Nevertheless, large language models provide a good initialisation for various sequence modelling tasks such as RL (<a href="http://can%20wikipedia%20help%20offline%20reinforcement%20learning/?">Reid et al., 2022</a>) or can be easily learned together with other sequence tasks (<a href="https://arxiv.org/pdf/2205.06175.pdf">Reed et al., 2022</a>). So simply the fact that something is hard to learn from text alone does not mean that large LMs will not be able to learn it.</p><p>Scaling effects may also disappear or appear differently when investigated in a larger-scale setting. For some of the Inverse Scaling Prize winners, <a href="https://arxiv.org/abs/2211.02011">Wei et al. (2022)</a> observe that the inverse scaling effect goes away when evaluated with 2x larger models and 5x more training compute. They also highlight the usefulness of chain-of-thought prompting to protect against inverse scaling. Inverse scaling thus depends on the specific training and test setting of a model.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ySnd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3e0fca6-6743-4515-b167-0ade34522635_2236x928.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ySnd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3e0fca6-6743-4515-b167-0ade34522635_2236x928.png 424w, https://substackcdn.com/image/fetch/$s_!ySnd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3e0fca6-6743-4515-b167-0ade34522635_2236x928.png 848w, https://substackcdn.com/image/fetch/$s_!ySnd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3e0fca6-6743-4515-b167-0ade34522635_2236x928.png 1272w, https://substackcdn.com/image/fetch/$s_!ySnd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3e0fca6-6743-4515-b167-0ade34522635_2236x928.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ySnd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3e0fca6-6743-4515-b167-0ade34522635_2236x928.png" width="1456" height="604" data-attrs="{&quot;src&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/a3e0fca6-6743-4515-b167-0ade34522635_2236x928.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:604,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:211779,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ySnd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3e0fca6-6743-4515-b167-0ade34522635_2236x928.png 424w, https://substackcdn.com/image/fetch/$s_!ySnd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3e0fca6-6743-4515-b167-0ade34522635_2236x928.png 848w, https://substackcdn.com/image/fetch/$s_!ySnd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3e0fca6-6743-4515-b167-0ade34522635_2236x928.png 1272w, https://substackcdn.com/image/fetch/$s_!ySnd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3e0fca6-6743-4515-b167-0ade34522635_2236x928.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">For some tasks where inverse scaling has been observed, the effect disappears with increasing model scale (<a href="https://arxiv.org/abs/2211.02011">Wei et al., 2022</a>).</figcaption></figure></div><p></p><h1>Image Generation is Heating Up &#128293;</h1><h2><strong>Imagen and Parti</strong></h2><p>After <a href="https://openai.com/dall-e-2/">DALL-E 2</a>, <a href="https://blog.google/technology/research/how-ai-creates-photorealistic-images-from-text/?utm_source=tw&amp;utm_medium=social&amp;utm_campaign=og&amp;utm_content=&amp;utm_term=">two new text-to-image models</a> have been released by Google, <a href="https://imagen.research.google/">Imagen</a> (pronounced "imagine") and <a href="https://parti.research.google/">Parti</a> (pronounced <a href="https://twitter.com/jasonbaldridge/status/1541847877315698689">"par-tee"</a>). Similar to DALL-E 2, Imagen is a <a href="https://newsletter.ruder.io/issues/palm-dall-e-2-chinchilla-chain-of-thought-prompting-values-and-culture-in-nlp-845878">diffusion model</a>. Parti, on the other hand, is auto-regressive and more like a standard encoder-decoder language model (LM). It learns to decode images into visual tokens, which are then converted by a <a href="https://ai.googleblog.com/2022/05/vector-quantized-image-modeling-with.html">ViT-VQGAN</a> model into an actual image.</p><p>As the <a href="http://%20these%20models%20have%20many%20limitations.%20for%20example%2C%20neither%20can%20reliably%20produce%20specific%20counts%20of%20objects%20%28e.g.xn--%20ten%20apples%29%2C%20nor%20place%20them%20correctly%20based%20on%20specific%20spatial%20descriptions%20%28e-ki74etd.g.xn--%20a%20red%20sphere%20to%20the%20left%20of%20a%20blue%20block%20with%20a%20yellow%20triangle%20on%20it%29-kh41e5f.%20also%2C%20as%20prompts%20become%20more%20complex%2C%20the%20models%20begin%20to%20falter%2C%20either%20missing%20details%20or%20introducing%20details%20that%20were%20not%20provided%20in%20the%20prompt.%20these%20behaviors%20are%20a%20result%20of%20several%20shortcomings%2C%20including%20lack%20of%20explicit%20training%20material%2C%20limited%20data%20representation%2C%20and%20lack%20of%203d%20awareness.%20we%20hope%20to%20address%20these%20gaps%20through%20broader%20representations%20and%20more%20effective%20integration%20into%20the%20text-to-image%20generation%20process./">blog post highlights</a>, limitations of current text-to-image models include their inability to count reliably, to follow precise spatial instructions, and to deal with complex prompts.</p><p>Given the potential of these models, it is not difficult to envision strategies to shore up each of these weaknesses. These can include collecting data that explicitly represents each phenomenon, using distant supervision with large amounts of data, and more explicit representations of number or spatial attributes in current models.</p><p>Another thing that can be challenging with current models is to generate natural text. For instance, DALL-E often produces gibberish as can be seen below.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BOpq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc6b53d-e308-4c6f-b5b1-8499a02213ee_740x709.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BOpq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc6b53d-e308-4c6f-b5b1-8499a02213ee_740x709.jpeg 424w, https://substackcdn.com/image/fetch/$s_!BOpq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc6b53d-e308-4c6f-b5b1-8499a02213ee_740x709.jpeg 848w, https://substackcdn.com/image/fetch/$s_!BOpq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc6b53d-e308-4c6f-b5b1-8499a02213ee_740x709.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!BOpq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc6b53d-e308-4c6f-b5b1-8499a02213ee_740x709.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BOpq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc6b53d-e308-4c6f-b5b1-8499a02213ee_740x709.jpeg" width="728" height="697.5027027027027" data-attrs="{&quot;src&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/bdc6b53d-e308-4c6f-b5b1-8499a02213ee_740x709.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:709,&quot;width&quot;:740,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:97514,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BOpq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc6b53d-e308-4c6f-b5b1-8499a02213ee_740x709.jpeg 424w, https://substackcdn.com/image/fetch/$s_!BOpq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc6b53d-e308-4c6f-b5b1-8499a02213ee_740x709.jpeg 848w, https://substackcdn.com/image/fetch/$s_!BOpq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc6b53d-e308-4c6f-b5b1-8499a02213ee_740x709.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!BOpq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc6b53d-e308-4c6f-b5b1-8499a02213ee_740x709.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">"Two farmers talking about vegetables, with subtitles"&#8212;as imagined by DALL-E 2.</figcaption></figure></div><p><a href="https://arxiv.org/abs/2206.00169">Daras et al. (2022)</a> fed some of the gibberish text it produced back to DALL-E 2 as a prompt, discovering that the model indeed generates images related to vegetables. They argue that many such terms are consistent across prompts and thus DALL-E 2 has a "secret" vocabulary (a "secret language" in the <a href="https://twitter.com/giannis_daras/status/1531693093040230402?s=20&amp;t=dqCtszWwzSjT0Kie8Ig1Rg">original Twitter thread</a>). In practice, such studies should be taken with a large grain of salt &#129474;. <a href="https://twitter.com/benjamin_hilton/status/1531780892972175361">Benjamin Hilton</a> highlights that most terms are not consistent across different prompts. So while terms like "Apoploe vesrreaitais" may kind of look like a biological name for birds and thus result in mostly bird-related generations, most other terms produced by DALL-E 2 are just noise&#8212;and definitely do not have the properties of a language such as grammar, morphology, etc.</p><p>After all, we know that models learning to play referential games often come up with degenerate communication protocols consisting of arbitrary symbols (<a href="https://openreview.net/forum?id=AUGBfDIV9rL">Chaabouni et al., 2022</a> is a nice recent study of the importance of scale for emergent communication). So I hope we go can turn down the hype and clickbaiting; let&#8217;s not go back to a time of newspapers reporting of <a href="https://www.techtimes.com/articles/212124/20170730/facebook-ai-invents-language-that-humans-cant-understand-system-shut-down-before-it-evolves-into-skynet.htm">AIs inventing a language that humans don't understand</a> &#128580;.</p><p>Larger variants of more recent models like Parti are also better at producing natural text&#8212;it seems this is another emergent ability of large multi-modal models.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vLua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8ec1a51-76d9-426a-8165-3c083e3e7f2b_740x204.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vLua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8ec1a51-76d9-426a-8165-3c083e3e7f2b_740x204.png 424w, https://substackcdn.com/image/fetch/$s_!vLua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8ec1a51-76d9-426a-8165-3c083e3e7f2b_740x204.png 848w, https://substackcdn.com/image/fetch/$s_!vLua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8ec1a51-76d9-426a-8165-3c083e3e7f2b_740x204.png 1272w, https://substackcdn.com/image/fetch/$s_!vLua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8ec1a51-76d9-426a-8165-3c083e3e7f2b_740x204.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vLua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8ec1a51-76d9-426a-8165-3c083e3e7f2b_740x204.png" width="740" height="204" data-attrs="{&quot;src&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/b8ec1a51-76d9-426a-8165-3c083e3e7f2b_740x204.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:204,&quot;width&quot;:740,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:214181,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vLua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8ec1a51-76d9-426a-8165-3c083e3e7f2b_740x204.png 424w, https://substackcdn.com/image/fetch/$s_!vLua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8ec1a51-76d9-426a-8165-3c083e3e7f2b_740x204.png 848w, https://substackcdn.com/image/fetch/$s_!vLua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8ec1a51-76d9-426a-8165-3c083e3e7f2b_740x204.png 1272w, https://substackcdn.com/image/fetch/$s_!vLua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8ec1a51-76d9-426a-8165-3c083e3e7f2b_740x204.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Emergent ability of producing natural text in images by Parti.</figcaption></figure></div><p>If you are envious of the high-quality generation of these closed large models, fear not. <a href="https://huggingface.co/spaces/dalle-mini/dalle-mini">DALL-E mini</a> is an open-source alternative created by <a href="https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini-Generate-images-from-any-text-prompt--VmlldzoyMDE4NDAy">Boris Dayma</a> that <a href="https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mini-Explained-with-Demo--Vmlldzo4NjIxODA">reproduces DALL-E</a> with a smaller architecture.</p><p>The model has become very popular lately with <a href="https://www.businessinsider.com/dall-e-mini">many websites</a> covering how to get around high traffic errors. My favourite part about DALL-E mini are a <a href="https://www.reddit.com/r/weirddalle/">reddit page</a> and <a href="https://twitter.com/weirddalle">Twitter account</a> that chronicle the often hilariously weird and surreal creations of the model.</p><div class="twitter-embed" data-attrs="{&quot;url&quot;:&quot;https://twitter.com/weirddalle/status/1534549407537963010&quot;,&quot;full_text&quot;:&quot;&quot;,&quot;username&quot;:&quot;weirddalle&quot;,&quot;name&quot;:&quot;Weird Dall-E Generations &#129484;&quot;,&quot;profile_image_url&quot;:&quot;&quot;,&quot;date&quot;:&quot;Wed Jun 08 14:54:23 +0000 2022&quot;,&quot;photos&quot;:[{&quot;img_url&quot;:&quot;https://pbs.substack.com/media/FUvQjUxXEAEthEL.jpg&quot;,&quot;link_url&quot;:&quot;https://t.co/DbLoe1s00c&quot;,&quot;alt_text&quot;:null}],&quot;quoted_tweet&quot;:{},&quot;reply_count&quot;:0,&quot;retweet_count&quot;:13409,&quot;like_count&quot;:164861,&quot;impression_count&quot;:0,&quot;expanded_url&quot;:{},&quot;video_url&quot;:null,&quot;belowTheFold&quot;:true}" data-component-name="Twitter2ToDOM"></div><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.ruder.io/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading NLP News! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[PaLM 🌴, DALL-E 2 👨‍🎨, Chinchilla 🐭, Chain-of-thought prompting ⛓💭✍️, Values and Culture in NLP 🏛]]></title><description><![CDATA[Hi all,This newsletter covers PaLM, DALL-E 2, and Chinchilla, chain-of-thought prompting, and the role of values and culture in NLP.This edition is somewhat delayed as I've been busy with planning a move (I'll be flying &#128747; to Germany tomorrow; say hi &#128075; if you're in Berlin) and exhausted by current events. I hope that you are all staying safe in these trying times &#127482;&#127462;.I really appreciate your feedback, so let me know what you&#160;love &#10084;&#65039; and hate &#128148; about this edition. Simply hit reply on the issue.Click here to view the newsletter in your browser.If you were referred by a friend, click&#160;here&#160;to subscribe. If you enjoyed this issue,&#160;give it a&#160;tweet&#160;&#128038;.]]></description><link>https://newsletter.ruder.io/p/palm-dall-e-2-chinchilla-chain-of-thought-prompting-values-and-culture-in-nlp-845878</link><guid isPermaLink="false">https://newsletter.ruder.io/p/palm-dall-e-2-chinchilla-chain-of-thought-prompting-values-and-culture-in-nlp-845878</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Sat, 16 Apr 2022 09:00:04 GMT</pubDate><enclosure url="https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi all,</p><p>This newsletter covers PaLM, DALL-E 2, and Chinchilla, chain-of-thought prompting, and the role of values and culture in NLP.</p><p>This edition is somewhat delayed as I've been busy with planning a move (I'll be flying &#128747; to Germany tomorrow; say hi &#128075; if you're in Berlin) and exhausted by current events. I hope that you are all staying safe in these trying times &#127482;&#127462;.</p><p>I really appreciate your feedback, so let me know what you&nbsp;love &#10084;&#65039; and hate &#128148; about this edition. Simply hit reply on the issue.</p><p>Click <strong><a href="https://newsletter.ruder.io/issues/palm-dall-e-2-chinchilla-chain-of-thought-prompting-values-and-culture-in-nlp-845878/3dcf8c68-12d0-4dbe-9f32-e683e9e2b2cc">here</a></strong> to view the newsletter in your browser.</p><p>If you were referred by a friend, click&nbsp;<a href="http://newsletter.ruder.io/?utm_campaign=NLP%20News&amp;utm_medium=email&amp;utm_source=Revue%20newsletter">here</a>&nbsp;to subscribe. If you enjoyed this issue,&nbsp;give it a&nbsp;<a href="https://twitter.com/intent/tweet?url=http%3A%2F%2Fnewsletter.ruder.io%2Fissues%2Fpalm-dall-e-2-chinchilla-chain-of-thought-prompting-values-and-culture-in-nlp-845878%2F3dcf8c68-12d0-4dbe-9f32-e683e9e2b2cc&amp;via=revue&amp;text=PaLM%20%F0%9F%8C%B4%2C%20DALL-E%202%20%F0%9F%91%A8%E2%80%8D%F0%9F%8E%A8%2C%20Chinchilla%20%F0%9F%90%AD%2C%20Chain-of-thought%20prompting%20%E2%9B%93%F0%9F%92%AD%E2%9C%8D%EF%B8%8F%2C%20Values%20and%20Culture%20in%20NLP%20%F0%9F%8F%9B%20by%20%40seb_ruder&amp;related=revue">tweet</a>&nbsp;&#128038;.</p><div><hr></div><h2>This Model Can Understand Your Jokes &#129322;</h2><p>The emergence of large pre-trained models has fundamentally changed the face and nature of progress in ML and NLP. The underlying methods have not changed dramatically; neural networks have already been pre-trained <a href="https://www.science.org/doi/pdf/10.1126/science.1127647?casa_token=jLTYgnADGqoAAAAA:VRlOXPfH5XsCgxieiLeXkrznIjKwKzvCgo3LEA_VaIr5-ZQU8v6cgGaXLZSYbMgNb6u0F8_54vsU">more than</a> <a href="https://proceedings.neurips.cc/paper/2006/file/5da713a690c067105aeb2fae32403405-Paper.pdf">15 years ago</a>. However, the recent scale of model size and data have enabled unprecedented&#8212;and indeed unexpected&#8212;capabilities.</p><p>Two recent models showcase the impressive progress in vision and NLP: OpenAI's <a href="https://openai.com/dall-e-2/">DALL-E 2</a> and Google's <a href="https://ai.googleblog.com/2022/04/pathways-language-model-palm-scaling-to.html">PaLM</a>. Both can be seen as the most recent milestone in a line of ever larger pre-trained models such as <a href="https://openai.com/blog/dall-e/">DALL-E</a>, <a href="https://arxiv.org/abs/1910.10683">T5</a> and <a href="https://arxiv.org/abs/2005.14165">GPT-3</a>, among others.</p><h2>DALL-E 2</h2><p><a href="https://cdn.openai.com/papers/dall-e-2.pdf?ref=hackernoon.com">DALL-E 2</a> consists of two components: a prior that generates a <a href="https://openai.com/blog/clip/">CLIP</a> image embedding based on a text description and a diffusion-based decoder that generates an image conditioned on an image embedding (see <a href="https://newsletter.ruder.io/issues/palm-dall-e-2-chinchilla-chain-of-thought-prompting-values-and-culture-in-nlp-845878/3dcf8c68-12d0-4dbe-9f32-e683e9e2b2cc">this overview</a> of CLIP and diffusion models). The <a href="https://cdn.openai.com/papers/dall-e-2.pdf?ref=hackernoon.com">paper</a>, unfortunately, is light on details regarding the composition and amount of training data. The resulting model produces more photorealistic and faithful images than its predecessor (see below).</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353 424w, https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353 848w, https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353 1272w, https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353 424w, https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353 848w, https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353 1272w, https://s3.amazonaws.com/revue/items/images/015/285/179/original/Screenshot_2022-04-15_at_11.39.10.png?1650019353 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Images generated by DALL-E 2.</figcaption></figure></div><p>While DALL-E 2 is able to generate impressive images, it still has weaknesses. As it relies on image&#8211;caption pairs for training, it may perform poorly when generating images that require more fine-grained visual reasoning such as <a href="https://twitter.com/pretendsmarts/status/1347824192557879297?s=20&amp;t=VkEKXaY2doa8RAEJYykwrw">counting</a>.</p><p>In NLP, there has been a debate on whether a language model trained only on unsupervised data can ever truly understand natural language (see <a href="https://newsletter.ruder.io/issues/ml-and-nlp-starter-toolkit-low-resource-nlp-toolkit-can-a-lm-understand-natural-language-the-next-generation-of-nlp-benchmarks-254211">this overview</a> and <a href="https://direct.mit.edu/tacl/article/doi/10.1162/tacl_a_00412/107385/Provable-Limitations-of-Acquiring-Meaning-from">this recent principled paper</a>). Multi-modal modals are grounded by definition. So are there any intrinsic limitations to the capabilities of language-and-vision models trained on image&#8211;caption alignments? One could argue that in order to learn truly multi-modal representations, a model must not only learn from depictions of the real world but <a href="https://arxiv.org/abs/2004.10151">must be able to interact with it</a>.</p><h2>PaLM</h2><p>On the language side, <a href="https://arxiv.org/abs/2204.02311">PaLM</a> is a 540B parameter decoder-only pre-trained Transformer model trained on multilingual&#8212;but heavily skewed towards English&#8212;data from the web as well as GitHub code. The model is evaluated in a few-shot setting on a battery of tasks where it generally outperforms the prior SOTA. On fine-tuning on SuperGLUE, the model handily outperforms the best decoder-only model and is competitive with encoder-decoder models (which generally perform better in such a fine-tuned setting).</p><h2>Joke explanations</h2><p>What I found most impressive, however, are some of the qualitative examples of model behaviour. For instance, the model is exceptionally good at explaining jokes. You can judge for yourself below. In each case, the model was prompted with just two example joke explanations and then had to generate its own.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/015/285/505/original/Screenshot_2022-04-15_at_12.07.14.png?1650020964" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/015/285/505/original/Screenshot_2022-04-15_at_12.07.14.png?1650020964 424w, https://s3.amazonaws.com/revue/items/images/015/285/505/original/Screenshot_2022-04-15_at_12.07.14.png?1650020964 848w, https://s3.amazonaws.com/revue/items/images/015/285/505/original/Screenshot_2022-04-15_at_12.07.14.png?1650020964 1272w, https://s3.amazonaws.com/revue/items/images/015/285/505/original/Screenshot_2022-04-15_at_12.07.14.png?1650020964 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/015/285/505/original/Screenshot_2022-04-15_at_12.07.14.png?1650020964" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/015/285/505/original/Screenshot_2022-04-15_at_12.07.14.png?1650020964&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/015/285/505/original/Screenshot_2022-04-15_at_12.07.14.png?1650020964 424w, https://s3.amazonaws.com/revue/items/images/015/285/505/original/Screenshot_2022-04-15_at_12.07.14.png?1650020964 848w, https://s3.amazonaws.com/revue/items/images/015/285/505/original/Screenshot_2022-04-15_at_12.07.14.png?1650020964 1272w, https://s3.amazonaws.com/revue/items/images/015/285/505/original/Screenshot_2022-04-15_at_12.07.14.png?1650020964 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Joke explanations by PaLM.</figcaption></figure></div><p>It will be interesting to see what this means for tasks such as sarcasm and irony detection, which have been mainstays of competitions such as <a href="https://codalab.lisn.upsaclay.fr/competitions/1340">SemEval</a>. I had previously considered these tasks to be still far out of reach of current model capabilities. Such anecdotal evidence naturally does not mean that these tasks are solved but that we may need more sophisticated and robust benchmarks to assess model performance.</p><p>Similarly, explaining jokes is not something that I would have expected current models to be able to do. Consequently, there may be an array of applications that have so far been infeasible where models might be able to add value. We can thus expect to see more work that explores how we can leverage such models for previous unexplored applications. For a large publicly available model to experiment with, check out <a href="https://arxiv.org/abs/2204.06745">GPT-NeoX-20B</a>.</p><p>Buoyed by these latest advances in NLP, there is a <a href="https://www.forbes.com/sites/robtoews/2022/03/27/a-wave-of-billion-dollar-language-ai-startups-is-coming/?sh=7a3c8e272b14">wave of new NLP startups</a> that tackle a diverse set of applications, from search to writing assistants, content moderation, and many more.</p><h2>Training Compute-Efficient LMs &#128045;</h2><p>While large language models are becoming more powerful, they are also becoming increasingly hard to use due to their huge size. In conjunction with scaling models, it is thus key to make advances in a) compression large models to smaller sizes and b) training more compute-efficient models to begin with.</p><p>Regarding the latter, researchers from DeepMind recently observed that <a href="https://arxiv.org/abs/2203.15556">current large language models are significantly under-trained</a>. They noticed that for the most compute-efficient training, when doubling the model size the number of training tokens should also be doubled. This is a much larger scaling rate than that predicted by <a href="https://arxiv.org/abs/2001.08361">previous scaling laws</a>. Their new 70B-parameter model, Chinchilla outperforms models of up to 530B parameters by training on much more data (1.4T vs 300B tokens).</p><p>Such an under-training phenomenon of large models is not entirely new. For <a href="https://arxiv.org/abs/1907.11692">RoBERTa</a>, the authors similarly observed that BERT was significantly under-trained and that longer training improves its performance.</p><p>Given the non-linear improvement and emergence of new capabilities with large model sizes, it will be key to investigate what is necessary to retain such impressive few-shot and reasoning capabilities at smaller model sizes.</p><p>Another direction I am excited about is the modularization of huge models: For most practical applications, not all capabilities of a huge model are truly relevant. How then can we isolate and compress a huge model's domain and task-specific knowledge in a small model that excels on the downstream task? Similarly, how can we efficiently leverage only the parts of the pre-trained model that are necessary for the downstream setting or bootstrap a strong small model using a large pre-trained model? For more thoughts on such a modular perspective of model development, check out Colin Raffel's <a href="https://colinraffel.com/blog/a-call-to-build-models-like-we-build-open-source-software.html">call to build open-source models like we build software</a>.</p><h2>Chain-of-Thought Prompting &#9939;&#128173;&#9997;&#65039;</h2><p>Despite the increasing power and capabilities of pre-trained models, the way we use and interact with them has not changed much. In-context prompting, pioneered by GPT-3 (see <a href="https://ruder.io/ml-highlights-2021/index.html#4prompting">this overview</a>) has been one of the most significant recent developments. However, we are still only scratching the surface of how to best extract information from pre-trained LMs and how to prime them for downstream tasks.</p><p>A method that enabled PaLM to perform particularly well on reasoning tasks is <em><a href="https://arxiv.org/abs/2201.11903">chain-of-thought prompting</a></em>. Rather than training a model to predict the answer, chain-of-thought prompting augments the prompt with an explanation of the reasoning steps to arrive at the answer as can be seen below.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/015/288/828/original/chain-of-thought_prompting.png?1650030117" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/015/288/828/original/chain-of-thought_prompting.png?1650030117 424w, https://s3.amazonaws.com/revue/items/images/015/288/828/original/chain-of-thought_prompting.png?1650030117 848w, https://s3.amazonaws.com/revue/items/images/015/288/828/original/chain-of-thought_prompting.png?1650030117 1272w, https://s3.amazonaws.com/revue/items/images/015/288/828/original/chain-of-thought_prompting.png?1650030117 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/015/288/828/original/chain-of-thought_prompting.png?1650030117" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/015/288/828/original/chain-of-thought_prompting.png?1650030117&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/015/288/828/original/chain-of-thought_prompting.png?1650030117 424w, https://s3.amazonaws.com/revue/items/images/015/288/828/original/chain-of-thought_prompting.png?1650030117 848w, https://s3.amazonaws.com/revue/items/images/015/288/828/original/chain-of-thought_prompting.png?1650030117 1272w, https://s3.amazonaws.com/revue/items/images/015/288/828/original/chain-of-thought_prompting.png?1650030117 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Standard prompting vs chain-of-thought prompting. The "chain of thought" is highlighted.</figcaption></figure></div><p>In a few-shot setting, these explanations can be manually written for a few examples. Prompted this way, the model learns to generate a similar explanation, which is particularly useful on more challenging reasoning problems.</p><h2>Related approaches</h2><p>Chain-of-thought prompting can be seen in line of several prior research areas. While explanations have been most commonly used to improve interpretability, <a href="https://aclanthology.org/P19-1487.pdf">Rajani et al. (2019)</a> train a model to automatically generate explanations during training and inference, achieving a new SOTA on a commonsense reasoning dataset.</p><p>In a similar vein, <a href="https://arxiv.org/abs/2112.00114">Nye et al. (2020)</a> train a model to write the intermediate computation steps of an arithmetic problem to a "scratchpad". For summarization, <a href="https://aclanthology.org/2021.tacl-1.88/">Narayan et al. (2021)</a> train a model to generate an entity chain (an ordered sequence of entities mentioned in the reference summary). At test time, the model first generates the entity chain before generating the summary.</p><p>There are other ways to improve learning with such intermediate outputs. <a href="https://arxiv.org/abs/2203.11171">Wang et al. (2022)</a> exploit the diversity of reasoning paths by sampling multiple chains of thought and then ensembling the final model predictions. As obtaining explanations for a large number of examples is expensive, <a href="https://arxiv.org/abs/2203.14465">Zelikman et al. (2022)</a> generate explanations for a large dataset by bootstrapping a model in the few-shot setting and only retaining explanations that lead to correct answers.</p><p>Using explanations, rationales or a description of reasoning steps works empirically but a more principled theory of how models leverage such rationales is still missing. In particular, it would be interesting to investigate to what extent a model's reasoning conforms to the reasoning steps preferred by humans (although the model can also be trained to perform more human-like reasoning, similar to <a href="https://openai.com/blog/instruction-following/">InstructGPT</a>).</p><h2>Interventions</h2><p>Beyond interpretability, generating an intermediate output enables the user to intervene on a model's predictions. <a href="https://aclanthology.org/2021.tacl-1.88/">Narayan et al. (2021)</a> demonstrate this by removing entities from the entity chain that were not seen in the original input, which improves the faithfulness of the generated summary. As a side-effect, such intermediate-output methods provide an interface and the potential to modulate and steer the predictions of otherwise black-box models. We can thus expect work focusing on whether such rationales truly explain model behaviour, similar to the <a href="https://aclanthology.org/P19-1282/">debate around</a> <a href="https://arxiv.org/abs/1902.10186">the explainability</a> <a href="https://arxiv.org/abs/1908.04626">of attention</a>.</p><h2>Outlook</h2><p>Overall, chain-of-thought prompting and related methods offer a glimpse of the untapped potential of current models. They also present an important and relatively compute-efficient research direction that can bring large improvements on top of state-of-the-art models. In this research, domain expertise is particularly important as it enables the development of strategies, reasoning steps, or alternative input methods that are particularly suited to an application. Prompts also do not need to be restricted to input&#8211;output pairs or explanations and can be much richer, including things to avoid, rules of thumb, positive or negative examples, etc as in the schema of <a href="https://arxiv.org/abs/2104.08773">Mishra et al. (2022)</a> below.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/015/289/344/original/instruction_schema.png?1650031891" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/015/289/344/original/instruction_schema.png?1650031891 424w, https://s3.amazonaws.com/revue/items/images/015/289/344/original/instruction_schema.png?1650031891 848w, https://s3.amazonaws.com/revue/items/images/015/289/344/original/instruction_schema.png?1650031891 1272w, https://s3.amazonaws.com/revue/items/images/015/289/344/original/instruction_schema.png?1650031891 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/015/289/344/original/instruction_schema.png?1650031891" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/015/289/344/original/instruction_schema.png?1650031891&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/015/289/344/original/instruction_schema.png?1650031891 424w, https://s3.amazonaws.com/revue/items/images/015/289/344/original/instruction_schema.png?1650031891 848w, https://s3.amazonaws.com/revue/items/images/015/289/344/original/instruction_schema.png?1650031891 1272w, https://s3.amazonaws.com/revue/items/images/015/289/344/original/instruction_schema.png?1650031891 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The schema used in the Natural Instructions dataset of Mishra et al. (2022).</figcaption></figure></div><h2>Values and Culture in NLP &#127963;</h2><p>One of the most important and far-reaching recent insights is that language models inherit the biases of the data they are trained on (see <a href="https://ruder.io/ml-highlights-2021/index.html#10bias">this overview</a>). Over time, our understanding of such biases has become more nuanced. Beyond generating toxic language when <a href="https://aclanthology.org/2020.findings-emnlp.301.pdf">conditioned with certain prompts</a>, recent work has turned to investigating a model's <em>ideology and values</em>.</p><p><a href="https://arxiv.org/abs/2201.10474">Gururangan et al. (2022)</a> aim to identify the language ideology encoded in GPT-3 by analyzing what type of language the quality filter used in GPT-3 is biased against. They replicate it and apply it to a corpus of U.S. high school newspapers (augmented with demographic information). They find that the filter favours text from authors who originate from regions with better educational attainment, urban centres, larger schools, and higher valued homes.</p><p>Looking at specific cultural values, <a href="https://arxiv.org/abs/2203.13722">Arora et al. (2022)</a> converted the questions of <a href="https://en.wikipedia.org/wiki/Hofstede%27s_cultural_dimensions_theory">Hofstede's cultural dimensions survey</a> and of the <a href="https://www.worldvaluessurvey.org/wvs.jsp">World Values survey</a> into prompts that were presented to multilingual language models. They find that the models exhibit differences in cultural values and that the values exhibited by the models are not in line with the values of the survey participants.</p><p><a href="https://arxiv.org/abs/2203.07785">Johnson et al. (2022)</a> investigate the values of GPT-3 by prompting it to summarize a range of culturally diverse texts. They examine the generated summaries and highlight problematic summaries and ones where the expressed values conflict with the original text.</p><p>These works demonstrate that beyond investigating bias related to specific lexical terms in current models, we also must be aware of the underlying values encoded in the model and expressed in the generated text. After all, we would not want our models to hold views that are outdated or disrespectful in certain cultural settings. However, the best way to investigate and robustly identify such values is still an open question.</p><p>The impact of culture in NLP goes beyond ideology and values. For a great overview of the cultural dimensions that are relevant in NLP, have a look at <a href="https://arxiv.org/abs/2203.10020">this survey</a>. The authors define four cultural dimensions of relevance: linguistic form and style (how things are expressed in language), common ground (shared knowledge based on which people reason and communicate), "aboutness" (what information is relevant or meaningful to people), and objectives or values (what people strive for).</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/015/290/223/original/cultural_dimensions.png?1650034527" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/015/290/223/original/cultural_dimensions.png?1650034527 424w, https://s3.amazonaws.com/revue/items/images/015/290/223/original/cultural_dimensions.png?1650034527 848w, https://s3.amazonaws.com/revue/items/images/015/290/223/original/cultural_dimensions.png?1650034527 1272w, https://s3.amazonaws.com/revue/items/images/015/290/223/original/cultural_dimensions.png?1650034527 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/015/290/223/original/cultural_dimensions.png?1650034527" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/015/290/223/original/cultural_dimensions.png?1650034527&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/015/290/223/original/cultural_dimensions.png?1650034527 424w, https://s3.amazonaws.com/revue/items/images/015/290/223/original/cultural_dimensions.png?1650034527 848w, https://s3.amazonaws.com/revue/items/images/015/290/223/original/cultural_dimensions.png?1650034527 1272w, https://s3.amazonaws.com/revue/items/images/015/290/223/original/cultural_dimensions.png?1650034527 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The role of culture in NLP.</figcaption></figure></div>]]></content:encoded></item><item><title><![CDATA[Highlights, new tasks & graph ML in 2021; Safer pre-trained models; Embeddings: Larger ≠ better]]></title><description><![CDATA[Hi all,I hope you've had a good start to the new year. This newsletter covers my and others' highlights of 2021. I also discuss recent pre-trained models that put more emphasis on safety and recent text similarity models where large is not always better. I really appreciate your feedback, so let me know what you&#160;love &#10084;&#65039; and hate &#128148; about this edition. Simply hit reply on the issue.Click here to view the newsletter in your browser.If you were referred by a friend, click&#160;here&#160;to subscribe. If you enjoyed this issue,&#160;give it a&#160;tweet&#160;&#128038;.]]></description><link>https://newsletter.ruder.io/p/highlights-new-tasks-graph-ml-in-2021-safer-pre-trained-models-embeddings-larger-better-900649</link><guid isPermaLink="false">https://newsletter.ruder.io/p/highlights-new-tasks-graph-ml-in-2021-safer-pre-trained-models-embeddings-larger-better-900649</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Mon, 31 Jan 2022 18:00:05 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!s2L_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0e246b0-416e-494b-915a-d5096d639de9_1746x988.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi all,</p><p>I hope you've had a good start to the new year. This newsletter covers <strong>my and others' highlights of 2021</strong>. I also discuss <strong>recent pre-trained models that put more emphasis on safety</strong> and recent <strong>text similarity models</strong> where <strong>large is not always better</strong>.</p><p>I really appreciate your feedback, so let me know what you&nbsp;love &#10084;&#65039; and hate &#128148; about this edition. Simply hit reply on the issue.</p><p>Click <strong><a href="https://newsletter.ruder.io/issues/highlights-new-tasks-graph-ml-in-2021-safer-pre-trained-models-embeddings-larger-better-900649/dc29f6b6-6586-4bc0-8c12-f7c27196e79b">here</a></strong> to view the newsletter in your browser.</p><p>If you were referred by a friend, click&nbsp;<a href="http://newsletter.ruder.io/?utm_campaign=NLP%20News&amp;utm_medium=email&amp;utm_source=Revue%20newsletter">here</a>&nbsp;to subscribe. If you enjoyed this issue,&nbsp;give it a&nbsp;<a href="https://twitter.com/intent/tweet?url=http%3A%2F%2Fnewsletter.ruder.io%2Fissues%2Fhighlights-new-tasks-graph-ml-in-2021-safer-pre-trained-models-embeddings-larger-better-900649%2Fdc29f6b6-6586-4bc0-8c12-f7c27196e79b&amp;via=revue&amp;text=Highlights%2C%20new%20tasks%20%26%20graph%20ML%20in%202021%3B%20Safer%20pre-trained%20models%3B%20Embeddings%3A%20Larger%20%E2%89%A0%20better%20by%20%40seb_ruder&amp;related=revue">tweet</a>&nbsp;&#128038;.</p><div><hr></div><h2>Looking back at 2021 &#128064;</h2><h2>ML and NLP Research Highlights of 2021 &#128161;</h2><p>I wrote up some of my research highlights in 2021 in <a href="https://ruder.io/ml-highlights-2021/">this post</a>. Overall, most of the trends I observed revolved around pre-trained models and their capabilities&#8212;how to <strong>train them more effectively</strong>, how to do <strong>few-shot learning</strong> with them, how to <strong>use them efficiently</strong>, how to <strong>evaluate</strong> them, using them for new applications such as <strong>program synthesis</strong>, etc. What were your highlights? You can share them by replying to the tweet below and I'll summarize them in the next newsletter.</p><div class="twitter-embed" data-attrs="{&quot;url&quot;:&quot;https://twitter.com/seb_ruder/status/1485710121816805377&quot;,&quot;full_text&quot;:&quot;ML and NLP Research Highlights of 2021\n\nThese are the research areas and papers I found most exciting &amp;amp; inspiring in 2021.\n\n&quot;,&quot;username&quot;:&quot;seb_ruder&quot;,&quot;name&quot;:&quot;Sebastian Ruder&quot;,&quot;profile_image_url&quot;:&quot;&quot;,&quot;date&quot;:&quot;Mon Jan 24 20:24:30 +0000 2022&quot;,&quot;photos&quot;:[],&quot;quoted_tweet&quot;:{},&quot;reply_count&quot;:0,&quot;retweet_count&quot;:425,&quot;like_count&quot;:1447,&quot;impression_count&quot;:0,&quot;expanded_url&quot;:{&quot;url&quot;:&quot;https://ruder.io/ml-highlights-2021/&quot;,&quot;image&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/f0e246b0-416e-494b-915a-d5096d639de9_1746x988.png&quot;,&quot;title&quot;:&quot;ML and NLP Research Highlights of 2021&quot;,&quot;description&quot;:&quot;This post summarizes progress across multiple impactful areas in ML and NLP in 2021.&quot;,&quot;domain&quot;:&quot;ruder.io&quot;},&quot;video_url&quot;:null,&quot;belowTheFold&quot;:true}" data-component-name="Twitter2ToDOM"></div><h2>New ML Tasks in 2021 &#128189;</h2><p>Another area I'm quite excited about is when ML is used to do new things. While such applications can be practically useful such as using AlphaFold 2.0 to accelerate the drug discovery process, I particularly enjoyed unconventional tasks or tasks provide a new perspective on existing research areas. Here are my favourites from 2021:</p><p><strong><a href="https://github.com/google/BIG-bench">BIG-bench</a></strong> contains a smorgasbord of diverse, sometimes quirky tasks for probing language models (LMs). <a href="https://github.com/google/BIG-bench/tree/main/bigbench/benchmark_tasks/checkmate_in_one">Predicting checkmate</a>? &#9989; <a href="https://github.com/google/BIG-bench/tree/main/bigbench/benchmark_tasks/emoji_movie">Guessing movies based on emojis</a>? &#9989; <a href="https://github.com/google/BIG-bench/tree/main/bigbench/benchmark_tasks/fantasy_reasoning">Reasoning in a fantasy world</a>? &#9989; The latter task includes examples like:</p><blockquote><p>As an amputee you experience phantom arm syndrome. Then one day you realize you can use it to punch ghosts. Your left arm is amputated but you still have your right arm.&nbsp;<em>Do you use your left arm to hit the late Elvis Presley to make him stop bothering you? </em>Answers:&nbsp;<strong>Yes </strong>/ No</p></blockquote><p>I don't know about you but I would prefer my ML model to resolve their problems without resorting to punching ghosts &#128123;.</p><p><strong>Cryptic crossword puzzles</strong> Solving cryptic crossword puzzles is a task that has attracted recent interest in the form of <a href="https://aclanthology.org/2021.emnlp-main.344/">two</a> <a href="https://openreview.net/forum?id=136ihvjd0sJ">datasets</a> and an associated <a href="https://github.com/google/BIG-bench/tree/main/bigbench/benchmark_tasks/cryptonite">BIG-bench task</a>. Crossword AIs have <a href="https://www.wired.com/story/crossword-ai-humans-way-with-words/">recently surpassed humans in a tournament</a> but cryptic clues are still very challenging for current models as these require both an understanding of semantics as well as wordplay (see below for an example from <a href="https://aclanthology.org/2021.emnlp-main.344/">Cryptonite</a>) that requires correctly identifying and resolving an anagram.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/013/838/150/original/cryptic_clue_example.png?1643539730" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/013/838/150/original/cryptic_clue_example.png?1643539730 424w, https://s3.amazonaws.com/revue/items/images/013/838/150/original/cryptic_clue_example.png?1643539730 848w, https://s3.amazonaws.com/revue/items/images/013/838/150/original/cryptic_clue_example.png?1643539730 1272w, https://s3.amazonaws.com/revue/items/images/013/838/150/original/cryptic_clue_example.png?1643539730 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/013/838/150/original/cryptic_clue_example.png?1643539730" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/013/838/150/original/cryptic_clue_example.png?1643539730&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/013/838/150/original/cryptic_clue_example.png?1643539730 424w, https://s3.amazonaws.com/revue/items/images/013/838/150/original/cryptic_clue_example.png?1643539730 848w, https://s3.amazonaws.com/revue/items/images/013/838/150/original/cryptic_clue_example.png?1643539730 1272w, https://s3.amazonaws.com/revue/items/images/013/838/150/original/cryptic_clue_example.png?1643539730 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p><strong>Reconstructing ancient texts</strong> The task of masked language modeling, filling in missing tokens in a text, lends itself directly to predicting missing tokens in the transliterated texts of ancient Akkadian clay tablets (<a href="https://aclanthology.org/2021.emnlp-main.384/">Lazar et al., 2021</a>). Such a setting is arguably more interesting than language modeling on the Penn Treebank&#8212;and trained models are practically useful by assisting experts in transcribing texts in extinct languages.</p><p><strong>Decontextualization</strong> is a new NLP task that requires rewriting an in-context sentence to be interpretable out of context (<a href="https://direct.mit.edu/tacl/article/doi/10.1162/tacl_a_00377/100685/Decontextualization-Making-Sentences-Stand-Alone">Choi et al., 2021</a>). This means dealing with various phenomena such as resolving coreferences and anaphora, adding relevant modifiers or necessary background information. Decontextualization is useful, for instance, in the context of question answering: instead of providing a sentence answer, which may be difficult to understand without the surrounding context, models can produce a sentence that stands on its own. See below for an example of how decontextualization looks like in practice.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/013/839/326/original/decontextualization_example.png?1643546811" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/013/839/326/original/decontextualization_example.png?1643546811 424w, https://s3.amazonaws.com/revue/items/images/013/839/326/original/decontextualization_example.png?1643546811 848w, https://s3.amazonaws.com/revue/items/images/013/839/326/original/decontextualization_example.png?1643546811 1272w, https://s3.amazonaws.com/revue/items/images/013/839/326/original/decontextualization_example.png?1643546811 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/013/839/326/original/decontextualization_example.png?1643546811" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/013/839/326/original/decontextualization_example.png?1643546811&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/013/839/326/original/decontextualization_example.png?1643546811 424w, https://s3.amazonaws.com/revue/items/images/013/839/326/original/decontextualization_example.png?1643546811 848w, https://s3.amazonaws.com/revue/items/images/013/839/326/original/decontextualization_example.png?1643546811 1272w, https://s3.amazonaws.com/revue/items/images/013/839/326/original/decontextualization_example.png?1643546811 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">An example decontextualization. The sentence to decontextualize is in gray.</figcaption></figure></div><p><strong>Text-based NP enrichment</strong> is a new information extraction task that focuses on extracting <em>all</em> relations (that are mediated by prepositions) between noun phrases in a text. NP enrichment unifies and complements many existing entity-related tasks such as relation extraction, semantic role labeling, entity linking, coreference resolution, etc. You can see how the annotation for this task looks like in the example below (<a href="https://arxiv.org/abs/2109.12085">Elazar et al., 2021</a>).</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/013/839/532/original/np_enrichment.png?1643548163" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/013/839/532/original/np_enrichment.png?1643548163 424w, https://s3.amazonaws.com/revue/items/images/013/839/532/original/np_enrichment.png?1643548163 848w, https://s3.amazonaws.com/revue/items/images/013/839/532/original/np_enrichment.png?1643548163 1272w, https://s3.amazonaws.com/revue/items/images/013/839/532/original/np_enrichment.png?1643548163 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/013/839/532/original/np_enrichment.png?1643548163" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/013/839/532/original/np_enrichment.png?1643548163&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/013/839/532/original/np_enrichment.png?1643548163 424w, https://s3.amazonaws.com/revue/items/images/013/839/532/original/np_enrichment.png?1643548163 848w, https://s3.amazonaws.com/revue/items/images/013/839/532/original/np_enrichment.png?1643548163 1272w, https://s3.amazonaws.com/revue/items/images/013/839/532/original/np_enrichment.png?1643548163 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Preposition-mediated relations between noun phrases in a text.</figcaption></figure></div><p>I am particularly excited by the newer tasks that explicitly go beyond core NLP tasks such as coreference resolution, which can be relatively narrow. Coreference resolution, for instance, is still useful to probe a model's reasoning abilities, e.g., as part of the <a href="https://en.wikipedia.org/wiki/Winograd_schema_challenge">Winograd schema</a> and later instantiations such as <a href="https://aclanthology.org/N18-2002/">Winogender</a> for gender bias and <a href="https://arxiv.org/abs/1907.10641">WinoGrande</a> for commonsense reasoning. However, as models become more powerful, we can apply them to a broader, more general set of problems, which may also be more practically useful.</p><p>I am constantly amazed by the emerging capabilities of models in ML and NLP and the new settings where they are applied and excited for the new things that we will be able to do this year.</p><h2><strong>Geometric &amp; Graph ML in 2021 &#128208;</strong></h2><p>Graph machine learning is one of the hottest emerging areas in ML. Graph ML methods are useful in a variety of domains, from modelling network data to molecules, interactions in physics, relations between entities, mathematical graphs, etc. Michael Bronstein and Peter Veli&#269;kovi&#263; interviewed experts in the area on their impressions of 2021 and predictions for 2022. <a href="https://towardsdatascience.com/predictions-and-hopes-for-geometric-graph-ml-in-2022-aa3b8b79f5cc">The article</a> is a great read for anyone who wants to get up to speed in this area. For a quick overview, you can check out <a href="https://towardsdatascience.com/predictions-and-hopes-for-geometric-graph-ml-in-2022-aa3b8b79f5cc#3eff">their take-home messages</a> highlighting, among others, the <strong>importance of message-passing</strong>&#8212;networks that update the hidden states of nodes based on information from adjacent nodes&#8212;, <strong>challenges of reasoning and generalisation</strong>, the <strong>combination of Transformers with graph neural networks</strong>, etc.</p><h2>Papers with Code 2021 &#128105;&#8205;&#128187;</h2><p>Papers with Code, one of the best resources for finding results, papers, and code in ML&#8212;which is also recently integrated into the ACL Anthology (see the code and data section at the bottom of a paper, such as <a href="https://aclanthology.org/P18-1031/">this one)</a>&#8212;highlights the top trending papers, libraries, and datasets of 2021. The <a href="https://arxiv.org/abs/2110.06635">most talked about paper</a> proposes a method to synthesize new views for an image from arbitrary camera angles, which captured people's attention with the below demo featuring impressive synthesized camera shots.</p><div id="youtube2-WJRyu1JUtVw" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;WJRyu1JUtVw&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/WJRyu1JUtVw?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>My 2021 &#128104;&#8205;&#128187;</strong></h2><p>My main threads of research in the past year were <strong>parameter efficiency</strong> (how can we make pre-trained models more efficient?), <strong>cross-lingual generalisation</strong> (how can multilingual models generalise better to under-represented languages?), and <strong>multilingual evaluation </strong>(see <a href="https://scholar.google.de/citations?hl=en&amp;user=8ONXPV8AAAAJ&amp;view_op=list_works&amp;sortby=pubdate">my Google Scholar</a> for the detailed publications). Some of the most fulfilling work was <strong>collaborating with passionate researchers</strong> such as from the <a href="https://www.masakhane.io/">Masakhane community</a> on <strong>building datasets in their own languages</strong>. I'm looking forward to doing more of this in 2022.</p><p>Like many, I had ups and downs. I've had less energy &#9889;&#65039; to do things outside of work, so have been less active on Twitter and only written blog posts and newsletters infrequently. I also regretted not being able to meet people from our community in person.</p><p>Overall, I'm hopeful for the new year and that things will slowly start going back to normal. I'm excited to write more in my spare time again and I'm looking forward to seeing many of you in person, at conferences or similar events.</p><h2>Safer Pre-trained Models &#128567;</h2><p>Prior work has found that pre-trained models are biased and can generate discriminatory or even toxic language. Ensuring safe responses is thus an important aspect of the development of such models. Recent models such as <a href="https://arxiv.org/abs/2201.08239">LaMDA</a>, <a href="https://cdn.openai.com/papers/Training_language_models_to_follow_instructions_with_human_feedback.pdf">InstructGPT</a>, and <a href="https://arxiv.org/abs/2112.11446">Gopher</a> developed by Google, OpenAI, and DeepMind respectively emphasize safety in their model evaluation and training. A common recipe is to <strong>fine-tune pre-trained models on data labeled with safety ratings by human annotators</strong>&#8212;using a reward model + RL or by training a detector and filtering out unsafe responses.</p><p>For <a href="https://arxiv.org/abs/2201.08239">LaMDA</a>, crowdworkers annotate model responses based on different safety criteria. The model is then fine-tuned both to generate dialogue responses as well as to predict the annotated safety labels. This multi-task setting is not only more efficient but also enables sharing information between the tasks. At test time, candidate responses where the model predicts a low safety rating are filtered out. The authors find that this fine-tuning setting significantly improves the safety of generated responses.</p><p>For <a href="https://cdn.openai.com/papers/Training_language_models_to_follow_instructions_with_human_feedback.pdf">InstructGPT</a>, GPT-3 is first fine-tuned on the demonstrations of annotators following instructions in a supervised setting. In a second step, raters rank multiple outputs of the fine-tuned model, which is used to train a reward model. Finally, the model is fine-tuned based on the output of the reward model using reinforcement learning. In an evaluation, the outputs of InstructGPT are significantly preferred over GPT-3's outputs while InstructGPT replaces GPT-3 in the API.</p><p>For <a href="https://arxiv.org/abs/2112.11446">Gopher</a>, the authors perform an extensive analysis of the toxicity and bias of the model. They find that larger models increase the toxicity of toxic input but do not amplify training data toxicity when unprompted. They also observe that large models are prone to bias against subgroups in a few-shot setting and that larger models are not able to overcome limitations in the coverage of dialects.</p><p>Overall, prior work as well as these recent efforts demonstrate that we cannot just pre-train models and expect them to produce safe or harmless responses. Instead, safety and inclusion need to be key design criteria that are included as part of the development of such models. This requires clearly enumerating and defining potential safety risks, collecting and annotating relevant data as well as explicitly training models to demonstrate safe behaviour. For recent reviews that highlight potential risks associated with language models, have a look <a href="https://ruder.io/ml-highlights-2021/#10bias">here</a>. I hope to see safety being considered as a design criterion and evaluation dimension in more work going forward.</p><h2>Embeddings: Larger &#8800; better &#127947;&#65039;&#8205;&#9792;&#65039;</h2><p>Nils Reimers <a href="https://medium.com/@nils_reimers/openai-gpt-3-text-embeddings-really-a-new-state-of-the-art-in-dense-text-embeddings-6571fe3ec9d9">analyzes embeddings</a> from OpenAI's recently released embeddings endpoint. OpenAI provides embeddings in different sizes, from 1,024&#8211;12,288 dimensions. He evaluates them on three downstream tasks&#8212;text similarity, text search, and code search.</p><p>He finds that the text similarity models perform much worse than state-of-the-art models such as <a href="https://huggingface.co/sentence-transformers/all-mpnet-base-v2">all-mpnet-base-v2</a> and <a href="https://huggingface.co/sentence-transformers/all-roberta-large-v1">all-roberta-large-v1</a>&#8212;<a href="https://arxiv.org/abs/2004.09297">MPNet</a> and <a href="https://arxiv.org/abs/1907.11692">RoBERTa</a> models respectively fine-tuned on 1B sentence pairs. They are also 6 points weaker than extremely small models with just 22M parameters that can run in a browser such as <a href="https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2">all-MiniLM-L6-v2</a>. On text search, they perform competitively but not quite at the level of the state of the art.</p><p>At the same time, due to their high dimensionality, the OpenAI embeddings are much slower than existing embedding models that have up to 768 dimensions and take up much more memory. He highlights that encoding the 21M passages of English Wikipedia in 384-dimensional embeddings requires about 16 GB while using 12,288 dimensions requires 516 GB of memory. Not only does retrieval using high-dimensional embeddings consume much more memory but it is also much slower than using smaller models.</p><p>Retrieval is also important for recent retrieval-augmented models such as <a href="https://arxiv.org/abs/2112.04426">Retro</a>, which retrieve from corpora of up to 2T tokens using frozen BERT representations (1,028 dimensions for BERT-large). Encoding such corpora with 12,288 dimensions would be prohibitive. Text similarity and retrieval-style tasks are one of the few settings these days where more parameters does not give you more bang for your buck; instead, for most realistic applications, low-dimensional performant embeddings are the way to go. Check out Nils' library <a href="https://www.sbert.net/">sentence-transformers</a> as well as the above models for efficient, powerful sentence representations.</p>]]></content:encoded></item><item><title><![CDATA[Pre-training + Massive Multi-tasking, Benchmarking in NLP, EMNLP primer, 🤗 NLP Course, ACL 2021 recap, ]]></title><description><![CDATA[Hi all, First off, some personal news: I've moved from DeepMind to Google Research this week. Because of this move, the past months have been quite busy. In light of this, I decided to pause the newsletter over the last couple of months. I plan to continue with it in a more sustainable manner.I'll be continuing to work on multilingual NLP, with a focus on under-represented languages, particularly those in Sub-Saharan Africa. On this note, if you are thinking of doing research in this area, I can't think of a better thing to do than apply for a funding opportunity via the Lacuna Fund. The call this time includes a mentorship program in collaboration with the amazing people from Masakhane.I really appreciate your feedback, so let me know what you&#160;love &#10084;&#65039; and hate &#128148; about this edition. Simply hit reply on the issue.Click here to view the newsletter in your browser.If you were referred by a friend, click&#160;here&#160;to subscribe. If you enjoyed this issue,&#160;give it a&#160;tweet&#160;&#128038;.]]></description><link>https://newsletter.ruder.io/p/pre-training-massive-multi-tasking-benchmarking-in-nlp-emnlp-primer-nlp-course-acl-2021-recap-709680</link><guid isPermaLink="false">https://newsletter.ruder.io/p/pre-training-massive-multi-tasking-benchmarking-in-nlp-emnlp-primer-nlp-course-acl-2021-recap-709680</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Sat, 06 Nov 2021 21:55:57 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/h_600,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi all,</p><p>First off, some personal news: I've moved from DeepMind to Google Research this week. Because of this move, the past months have been quite busy. In light of this, I decided to pause the newsletter over the last couple of months. I plan to continue with it in a more sustainable manner.</p><p>I'll be continuing to work on multilingual NLP, with a focus on under-represented languages, particularly those in Sub-Saharan Africa. On this note, if you are thinking of doing research in this area, I can't think of a better thing to do than <strong>apply for a funding opportunity</strong> via the <a href="https://lacunafund.org/apply/">Lacuna Fund</a>. The call this time includes a mentorship program in collaboration with the amazing people from <a href="https://www.masakhane.io/">Masakhane</a>.</p><p>I really appreciate your feedback, so let me know what you&nbsp;love &#10084;&#65039; and hate &#128148; about this edition. Simply hit reply on the issue.</p><p>Click <strong><a href="https://newsletter.ruder.io/issues/pre-training-massive-multi-tasking-709680/05e59718-2554-4a0c-84d2-4e1572a020a2">here</a></strong> to view the newsletter in your browser.</p><p>If you were referred by a friend, click&nbsp;<a href="http://newsletter.ruder.io/?utm_campaign=NLP%20News&amp;utm_medium=email&amp;utm_source=Revue%20newsletter">here</a>&nbsp;to subscribe. If you enjoyed this issue,&nbsp;give it a&nbsp;<a href="https://twitter.com/intent/tweet?url=http%3A%2F%2Fnewsletter.ruder.io%2Fissues%2Fpre-training-massive-multi-tasking-benchmarking-in-nlp-emnlp-primer-nlp-course-acl-2021-recap-709680%2F05e59718-2554-4a0c-84d2-4e1572a020a2&amp;via=revue&amp;text=Pre-training%20%2B%20Massive%20Multi-tasking%2C%20%20Benchmarking%20in%20NLP%2C%20EMNLP%20primer%2C%20%F0%9F%A4%97%20NLP%20Course%2C%20ACL%202021%20recap%2C%20%20by%20%40seb_ruder&amp;related=revue">tweet</a>&nbsp;&#128038;.</p><div><hr></div><h2>Pre-training + Massive Multi-tasking &#128145;</h2><p>Multi-task learning (MTL), training a model on several tasks at once and sharing information is a general method that is fundamental to training neural networks. <a href="https://www.cs.cornell.edu/~caruana/mlj97.pdf">Rich Caruana's 1997 paper</a> is one of the best introductions to this topic and as relevant today as it was back then. For more recent overviews, you can check out <a href="https://arxiv.org/abs/1706.05098">my survey from 2017</a> or a <a href="https://arxiv.org/abs/2009.09796">survey from 2020</a> that I enjoyed.</p><p>Research in multi-task learning has long shown that models trained on many tasks <strong>learn representations that generalize better to new ones</strong>. A common problem in multi-task learning, however, is minimizing negative transfer, i.e. how to make sure that tasks that are dissimilar do not hurt each other.</p><p>In recent years despite much work on alternative training objectives, the NLP community has gravitated to a single pre-training objective<em> to rule them all</em>, <strong>masked language modelling (MLM)</strong>. Much recent work has focused on ways to adapt and improve it (e.g., <a href="https://openreview.net/forum?id=3Aoft6NWFej">Levine et al., 2021</a>). Even the next-sentence-prediction objective used in BERT has become slowly phased out (<a href="https://aclanthology.org/2020.emnlp-main.403/">Aroca-Ouellette &amp; Rudzicz, 2020</a>).</p><p>Recently, there has been a flurry of papers that show not only that multi-task learning helps pre-trained models, but that <strong>gains are larger when more tasks are used</strong>. Such massive multi-task learning settings cover up to around 100 tasks, going beyond earlier work that covered around 50 tasks (<a href="https://arxiv.org/abs/2101.11038">Aghajanyan et al., 2021</a>).</p><p>A key reason for this convergence of papers is that <strong>multi-task learning is much easier with recent models</strong>, even across many tasks. This is due to the fact that many recent models such as T5 and GPT-3 <strong>use a text-to-text format</strong>. Gone are the days of hand-engineered task-specific loss functions for multi-task learning. Instead, each task only needs to be expressed in a suitable text-to-text format and models will be able to learn from it, without any changes to the underlying model.</p><p>The newly proposed approaches differ in terms of <em>how</em> and <em>when</em> multi-task learning is applied. One choice is <strong>fine-tuning an existing pre-trained model</strong> on a collection of multiple tasks, i.e. <a href="https://ruder.io/recent-advances-lm-fine-tuning/#behavioural-fine-tuning">behavioural fine-tuning</a>. This is done by T0 (<a href="https://arxiv.org/abs/2110.08207">Sanh et al., 2021</a>), one of the first outcomes of the <a href="https://bigscience.huggingface.co/">BigScience workshop</a>, using T5 and FLAN (<a href="https://arxiv.org/abs/2109.01652">Wei et al., 2021</a>) using a GPT-3-like pre-trained model. Both papers describe a unified template and instruction format into which they convert existing datasets. BigScience open-sources their collection of prompts <a href="https://github.com/bigscience-workshop/promptsource">here</a>. Both papers <strong>report large improvements in terms of zero-shot and few-shot performance</strong> compared to state-of-the-art models like T5 and GPT-3.</p><p><a href="https://arxiv.org/abs/2110.15943">Min et al. (2021)</a> propose a different fine-tuning setting that optimizes for <em>in-context learning</em>: instead of fine-tuning a model on examples of a task directly, they provide the concatenation of <em>k+1</em> examples to a model as input <em>x_1, y_1, ..., x_k, y_k, x_{k+1}</em> and train the model to predict the label of the <em>k+1</em>-th example, <em>y_{k+1}</em>. They similarly report improvements in zero-shot transfer.</p><p>In contrast to the previous approaches, ExT5 (<a href="https://openreview.net/pdf?id=Vzh1BFUCiIX">Anonymous et al., 2021</a>) <em>pre-trains</em> a model on a large collection of tasks. They observe that using multiple tasks during pre-training is better than during fine-tuning and that <strong>multi-task pre-training combined with MLM is significantly more sample-efficient</strong> than just using MLM (see below).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Knjt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Knjt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png 424w, https://substackcdn.com/image/fetch/$s_!Knjt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png 848w, https://substackcdn.com/image/fetch/$s_!Knjt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png 1272w, https://substackcdn.com/image/fetch/$s_!Knjt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Knjt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png" width="1159" height="549" data-attrs="{&quot;src&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:549,&quot;width&quot;:1159,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Knjt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png 424w, https://substackcdn.com/image/fetch/$s_!Knjt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png 848w, https://substackcdn.com/image/fetch/$s_!Knjt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png 1272w, https://substackcdn.com/image/fetch/$s_!Knjt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F46797142-67ff-4a9d-8cd0-da41b7e2dcca_1159x549.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">SuperGLUE score of ExT5-LARGE vs T5-LARGE as a function of number of pre-training steps</figcaption></figure></div><p>On the whole, these papers highlight the <strong>benefit of combining </strong><em><strong>self-supervised</strong></em><strong> pre-training with </strong><em><strong>supervised</strong></em><strong> multi-task learning</strong>. While multi-task fine-tuned models were always somewhat inferior to single-task models on small task collections such as GLUE&#8212;with a few exceptions (<a href="https://aclanthology.org/P19-1441/">Liu et al., 2019</a>; <a href="https://aclanthology.org/P19-1595/">Clark et al., 2019</a>)&#8212;<strong>multi-task models may soon hold state-of-the-art results on many benchmarks</strong>. Given the availability and open-source nature of datasets in a unified format, we can imagine a <strong>virtuous cycle</strong> where newly created high-quality datasets are used to train more powerful models on increasingly diverse task collections, which could then be used in-the-loop to create more challenging datasets.</p><p>In light of the increasingly multi-task nature of such models, <strong>what then does it mean to do zero-shot learning</strong>? In current training setups, datasets from certain tasks such as NLI are excluded from training in order to ensure a fair zero-shot scenario at test time. As open-source multi-task models trained on many existing tasks become more common, it will be increasingly difficult to guarantee a setting where a model has not seen examples of a similar task. In this context, few-shot learning or the full supervised setting may become the preferred evaluation paradigms.</p><h2>Benchmarking in NLP &#129351;</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UrXv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F89ad1f90-3cb6-41b6-b307-c58c528b8c49_1724x1232.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UrXv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F89ad1f90-3cb6-41b6-b307-c58c528b8c49_1724x1232.png 424w, https://substackcdn.com/image/fetch/$s_!UrXv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F89ad1f90-3cb6-41b6-b307-c58c528b8c49_1724x1232.png 848w, https://substackcdn.com/image/fetch/$s_!UrXv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F89ad1f90-3cb6-41b6-b307-c58c528b8c49_1724x1232.png 1272w, https://substackcdn.com/image/fetch/$s_!UrXv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F89ad1f90-3cb6-41b6-b307-c58c528b8c49_1724x1232.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UrXv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F89ad1f90-3cb6-41b6-b307-c58c528b8c49_1724x1232.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/89ad1f90-3cb6-41b6-b307-c58c528b8c49_1724x1232.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!UrXv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F89ad1f90-3cb6-41b6-b307-c58c528b8c49_1724x1232.png 424w, https://substackcdn.com/image/fetch/$s_!UrXv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F89ad1f90-3cb6-41b6-b307-c58c528b8c49_1724x1232.png 848w, https://substackcdn.com/image/fetch/$s_!UrXv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F89ad1f90-3cb6-41b6-b307-c58c528b8c49_1724x1232.png 1272w, https://substackcdn.com/image/fetch/$s_!UrXv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F89ad1f90-3cb6-41b6-b307-c58c528b8c49_1724x1232.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">&#8212;Most ML and NLP models, probably</figcaption></figure></div><p>Many talks and papers at ACL 2021 made reference to the current state of NLP benchmarking, which has seen existing benchmarks largely outpaced by rapidly improving pre-trained models, in spite of such models still being far away from true human-level natural language understanding.</p><p>My favourite resources on this topic from the conference were:</p><ul><li><p>Chris Pott's keynote on <a href="https://www.youtube.com/watch?v=t_A36DDcG_0">Reliable characterizations of NLP systems as a social responsibility</a>.</p></li><li><p><a href="https://www.youtube.com/watch?v=FXCSWvIsdEE">Rada Mihalcea's presidential address</a> where she emphasises evaluation beyond accuracy.</p></li><li><p>Samuel Bowman's and George Dahl's position paper asking "<a href="https://aclanthology.org/2021.naacl-main.385.pdf">What Will it Take to Fix Benchmarking in Natural Language Understanding?</a>"</p></li></ul><p>I've also written a <a href="https://ruder.io/nlp-benchmarking/">longer blog post</a> that provides a broader overview of different perspectives, challenges, and potential solutions to improve benchmarking in NLP.</p><h2>EMNLP 2021 primer &#127965;</h2><p>EMNLP papers are now available in the <a href="https://aclanthology.org/events/emnlp-2021/">ACL anthology</a>. I'll be attending the conference virtually. I'm particularly looking forward to the virtual poster sessions (as these feel closest to an in-person conference experience).</p><p>In terms of the conference program, I was particularly excited to see <strong>Visually Grounded Reasoning across Languages and Cultures</strong> (<a href="https://aclanthology.org/2021.emnlp-main.818/">Liu et al., 2021</a>) being awarded the Best paper award. The paper highlights biases in the concepts and images in ImageNet and proposes a new multilingual dataset for Multicultural Reasoning over Vision and Language (<a href="https://marvl-challenge.github.io/">MaRVL</a>; see below for examples). We need more work that creates such high-quality datasets in a culturally diverse setting.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!U4Kj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F1353c808-15ed-4711-9228-7dd99c1136b5_915x1077.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!U4Kj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F1353c808-15ed-4711-9228-7dd99c1136b5_915x1077.png 424w, https://substackcdn.com/image/fetch/$s_!U4Kj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F1353c808-15ed-4711-9228-7dd99c1136b5_915x1077.png 848w, https://substackcdn.com/image/fetch/$s_!U4Kj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F1353c808-15ed-4711-9228-7dd99c1136b5_915x1077.png 1272w, https://substackcdn.com/image/fetch/$s_!U4Kj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F1353c808-15ed-4711-9228-7dd99c1136b5_915x1077.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!U4Kj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F1353c808-15ed-4711-9228-7dd99c1136b5_915x1077.png" width="915" height="1077" data-attrs="{&quot;src&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/1353c808-15ed-4711-9228-7dd99c1136b5_915x1077.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1077,&quot;width&quot;:915,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!U4Kj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F1353c808-15ed-4711-9228-7dd99c1136b5_915x1077.png 424w, https://substackcdn.com/image/fetch/$s_!U4Kj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F1353c808-15ed-4711-9228-7dd99c1136b5_915x1077.png 848w, https://substackcdn.com/image/fetch/$s_!U4Kj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F1353c808-15ed-4711-9228-7dd99c1136b5_915x1077.png 1272w, https://substackcdn.com/image/fetch/$s_!U4Kj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F1353c808-15ed-4711-9228-7dd99c1136b5_915x1077.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Examples from MaRVL in Tamil (a) and Swahili (b).</figcaption></figure></div><p>If this is your first time at a conference, I would recommend you to prioritize meeting and connecting with people. Approach others and introduce yourself (whether in-person or at the virtual GatherTown). Go to posters, both those which are well-attended and those where only the presenter is present; the latter often make for a more insightful and stimulating discussion. Above all else, have fun (we all need it after the last two years). If you are presenting something at the conference or just want to say hi, feel free to send me a message.</p><p>You can also find me doing the following things during the conference:</p><ul><li><p>November 9: presenting a poster (<a href="https://underline.io/events/192/posters?eventSessionId=8269">poster session link</a>) and talk (<a href="https://underline.io/events/192/sessions?eventSessionId=7831">oral session link</a>) for <strong>XTREME-R</strong> (<a href="https://aclanthology.org/2021.emnlp-main.802/">paper link</a>)</p></li><li><p>November 10: presenting a keynote on multilingual evaluation at the <strong>Eval4NLP workshop</strong> (<a href="https://underline.io/events/192/sessions?eventSessionId=7851">conference link</a>, <a href="https://eval4nlp.github.io/">external link</a>)</p></li><li><p>November 11: co-presenting a hybrid tutorial on <strong>Multi-Domain Multilingual Question Answering</strong> (<a href="https://underline.io/events/192/sessions?eventSessionId=7844">conference session link</a>; <a href="https://docs.google.com/presentation/d/1cML2FBrevPF0BtQu5QH_BL3dUbfn0egYyhTbOgwqlbA/edit?usp=sharing">slides link</a>) as well as co-hosting <strong>The 1st Workshop on Multi-lingual Representation Learning</strong> (<a href="https://underline.io/events/192/sessions?eventSessionId=7861">conference link</a>; <a href="https://sites.google.com/corp/view/mrl-2021/home">external link</a>)</p></li></ul><p>Also check out the following EMNLP 2021 together with collaborators:</p><ul><li><p><strong><a href="https://aclanthology.org/2021.emnlp-main.699/">IndoNLG: Benchmark and Resources for Evaluating Indonesian Natural Language Generation</a></strong>: The first benchmark (and new models) for NLG in Indonesian.</p></li><li><p><strong><a href="https://aclanthology.org/2021.emnlp-main.800/">UNKs Everywhere: Adapting Multilingual Language Models to New Scripts</a></strong>: New methods + a system evaluation to adapt pre-trained models to unseen scripts.</p></li><li><p><strong><a href="https://aclanthology.org/2021.findings-emnlp.63/">Efficient Test Time Adapter Ensembling for Low-resource Language Varieties</a></strong> (Findings): A simple ensembling-based method to use existing language adapters for adapting to unseen languages and language varieties.</p></li><li><p><strong><a href="https://aclanthology.org/2021.findings-emnlp.410/">MAD-G: Multilingual Adapter Generation for Efficient Cross-Lingual Transfer</a></strong> (Findings): A method based on hyper-networks to generate language adapters for cross-lingual transfer.</p></li></ul><h2>Upcoming NLP course by Hugging Face &#129303;</h2><p>One thing that is great about doing NLP these days is that there are a lot of resources to help you get started as well as a lot of tooling and infrastructure to work easily with state-of-the-art models.</p><p>A nice resource for learning about using NLP is the <a href="https://huggingface.co/course/chapter0?fw=pt">NLP course by Hugging Face</a>. Four chapters are currently available, with more to be released by November 16. The new chapters discuss how to create datasets and tokenizers as well as how to deal with many standard NLP settings, such as fine-tuning a model for translation, summarization, or question answering (see below).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2853!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fda85eaa9-036e-4076-88cc-4df81b5a0e18_1405x908.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2853!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fda85eaa9-036e-4076-88cc-4df81b5a0e18_1405x908.png 424w, https://substackcdn.com/image/fetch/$s_!2853!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fda85eaa9-036e-4076-88cc-4df81b5a0e18_1405x908.png 848w, https://substackcdn.com/image/fetch/$s_!2853!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fda85eaa9-036e-4076-88cc-4df81b5a0e18_1405x908.png 1272w, https://substackcdn.com/image/fetch/$s_!2853!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fda85eaa9-036e-4076-88cc-4df81b5a0e18_1405x908.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2853!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fda85eaa9-036e-4076-88cc-4df81b5a0e18_1405x908.png" width="1405" height="908" data-attrs="{&quot;src&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/da85eaa9-036e-4076-88cc-4df81b5a0e18_1405x908.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:908,&quot;width&quot;:1405,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!2853!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fda85eaa9-036e-4076-88cc-4df81b5a0e18_1405x908.png 424w, https://substackcdn.com/image/fetch/$s_!2853!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fda85eaa9-036e-4076-88cc-4df81b5a0e18_1405x908.png 848w, https://substackcdn.com/image/fetch/$s_!2853!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fda85eaa9-036e-4076-88cc-4df81b5a0e18_1405x908.png 1272w, https://substackcdn.com/image/fetch/$s_!2853!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2Fda85eaa9-036e-4076-88cc-4df81b5a0e18_1405x908.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The Question answering chapter of the NLP course</figcaption></figure></div><h2>ACL 2021 recap &#127963;</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5zkB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6b76cf-d0c4-4cb8-95cc-63a53cd68522_1232x618.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5zkB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6b76cf-d0c4-4cb8-95cc-63a53cd68522_1232x618.png 424w, https://substackcdn.com/image/fetch/$s_!5zkB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6b76cf-d0c4-4cb8-95cc-63a53cd68522_1232x618.png 848w, https://substackcdn.com/image/fetch/$s_!5zkB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6b76cf-d0c4-4cb8-95cc-63a53cd68522_1232x618.png 1272w, https://substackcdn.com/image/fetch/$s_!5zkB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6b76cf-d0c4-4cb8-95cc-63a53cd68522_1232x618.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5zkB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6b76cf-d0c4-4cb8-95cc-63a53cd68522_1232x618.png" width="1232" height="618" data-attrs="{&quot;src&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/2f6b76cf-d0c4-4cb8-95cc-63a53cd68522_1232x618.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:618,&quot;width&quot;:1232,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!5zkB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6b76cf-d0c4-4cb8-95cc-63a53cd68522_1232x618.png 424w, https://substackcdn.com/image/fetch/$s_!5zkB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6b76cf-d0c4-4cb8-95cc-63a53cd68522_1232x618.png 848w, https://substackcdn.com/image/fetch/$s_!5zkB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6b76cf-d0c4-4cb8-95cc-63a53cd68522_1232x618.png 1272w, https://substackcdn.com/image/fetch/$s_!5zkB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F2f6b76cf-d0c4-4cb8-95cc-63a53cd68522_1232x618.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">"Hot topics" in ACL 2021 papers compared to ACL 2018</figcaption></figure></div><p>This should come as no surprise but it's still interesting to see that <strong>among the 14 "hot" topics of 2021 (see above) were 5 pre-trained models</strong> (BERT, RoBERTa, BART, GPT-2, XLM-R) and 1 general "Language models" topic. These models are essentially all variants of the same Transformer architecture. This serves as a useful reminder that the community may be overfitting to a particular setting and that it may be worthwhile to look beyond the standard Transformer model (see <a href="https://newsletter.ruder.io/issues/github-copilot-the-perceiver-beyond-the-transformer-data-augmentation-nl-augmenter-research-communication-527358">my recent newsletter</a> for some inspiration).</p><p>It seems everyone is exhausted by virtual conferences at this point as I wasn't able to find any write-ups of people's highlights of ACL 2021, in contrast to past conferences (I also didn't manage to finish mine).</p><p>If you're attending EMNLP 2021, I hope you'll share your highlights, experience, and insights with the community.</p>]]></content:encoded></item><item><title><![CDATA[ICML round-up, Open collaboration, CLIP art, Internet augmentation, New GLUE-style benchmarks]]></title><description><![CDATA[Hi all,This newsletter covers some of my favourite papers from ICML 2021, a discussion of open collaboration, art generated by the CLIP model, how to leverage information from the Internet in your models, and new benchmarks in the style of GLUE.FYI, I'll be at ACL 2021 virtually this week. Ping me on Gathertown or send me an email if you would like to chat. I'm co-author on two papers on parameter-efficient multi-task learning and monolingual vs multilingual models, which will be presented by the first authors on 12 pm, August 2 and 11 am, August 3 respectively.I really appreciate your feedback, so let me know what you&#160;love &#10084;&#65039; and hate &#128148; about this edition. Simply hit reply on the issue.Click here to view the newsletter in your browser.If you were referred by a friend, click&#160;here&#160;to subscribe. If you enjoyed this issue,&#160;give it a&#160;tweet&#160;&#128038;.]]></description><link>https://newsletter.ruder.io/p/icml-round-up-open-collaboration-clip-art-internet-augmentation-new-glue-style-benchmarks-692828</link><guid isPermaLink="false">https://newsletter.ruder.io/p/icml-round-up-open-collaboration-clip-art-internet-augmentation-new-glue-style-benchmarks-692828</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Mon, 02 Aug 2021 08:30:01 GMT</pubDate><enclosure url="https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi all,</p><p>This newsletter covers some of my favourite papers from ICML 2021, a discussion of open collaboration, art generated by the CLIP model, how to leverage information from the Internet in your models, and new benchmarks in the style of GLUE.</p><p>FYI, I'll be at <a href="https://2021.aclweb.org/">ACL 2021</a> virtually this week. Ping me on Gathertown or send me an email if you would like to chat. I'm co-author on two papers on <a href="https://aclanthology.org/2021.acl-long.47/?utm_campaign=NLP%20News&amp;utm_medium=email&amp;utm_source=Revue%20newsletter">parameter-efficient multi-task learning</a> and <a href="https://aclanthology.org/2021.acl-long.243/?utm_campaign=NLP%20News&amp;utm_medium=email&amp;utm_source=Revue%20newsletter">monolingual vs multilingual models</a>, which will be presented by the first authors on 12 pm, August 2 and 11 am, August 3 respectively.</p><p>I really appreciate your feedback, so let me know what you&nbsp;love &#10084;&#65039; and hate &#128148; about this edition. Simply hit reply on the issue.</p><p><em>Click <strong><a href="https://newsletter.ruder.io/issues/icml-round-up-open-collaboration-clip-art-internet-augmentation-new-glue-style-benchmarks-692828/ffafaca9-360c-41b2-a31e-7bf8eabc5619">here</a></strong> to view the newsletter in your browser.</em></p><p>If you were referred by a friend, click&nbsp;<a href="http://newsletter.ruder.io/?utm_campaign=NLP%20News&amp;utm_medium=email&amp;utm_source=Revue%20newsletter">here</a>&nbsp;to subscribe. If you enjoyed this issue,&nbsp;give it a&nbsp;<a href="https://twitter.com/intent/tweet?url=http%3A%2F%2Fnewsletter.ruder.io%2Fissues%2Ficml-round-up-open-collaboration-clip-art-internet-augmentation-new-glue-style-benchmarks-692828%2Fffafaca9-360c-41b2-a31e-7bf8eabc5619&amp;via=revue&amp;text=ICML%20round-up%2C%20Open%20collaboration%2C%20CLIP%20art%2C%20Internet%20augmentation%2C%20New%20GLUE-style%20benchmarks%20by%20%40seb_ruder&amp;related=revue">tweet</a>&nbsp;&#128038;.</p><div><hr></div><h2>ICML round-up &#128209;</h2><p><strong><a href="http://proceedings.mlr.press/v139/lin21b/lin21b.pdf">Straight to the Gradient: Learning to Use Novel Tokens for Neural Text Generation</a></strong> Neural generative models, despite their popularity, are known to suffer from some deficiencies, such as a tendency to generate frequent tokens. Popular methods to address this such as top-<em>k</em> sampling <a href="https://aclanthology.org/P18-1082.pdf">(Fan et al., 2018)</a> or nucleus sampling (<a href="http://The Curious Case of Neural Text Degeneration">Holtzman et al., 2020</a>) focus on <em>decoding</em>. This paper proposes ScaleGrad, which re-scales the token probabilities during training to encourage the model to focus on novel tokens, i.e. ones that have not been generated before. ScaleGrad seems to improve performance on some open-ended as well as directed generation tasks. Of course, just focusing on novel tokens may be too simplistic. <strong>Overall, modifying the loss function with regard to particular sets of tokens may be a useful way to inject additional inductive biases into a model, such as which entities or attributes to focus on.</strong></p><p><strong><a href="http://proceedings.mlr.press/v139/price21a/price21a.pdf">Dense for the Price of Sparse: Improved Performance of Sparsely Initialized Networks via a Subspace Offset</a></strong> This paper nicely highlights the <em>competing</em> <em>priorities</em> when training sparse networks: The current trend of identifying 'lottery tickets', i.e. sparse subnetworks that can be trained on their own from scratch and which perform similarly to full networks, is motivated by computational efficiency. However, such methods require computing a score for all parameters in the full model to determine whether they should be pruned. <strong>It is thus still necessary to store and compute with the full model on device.</strong> To reduce on-device storage costs, the authors propose a network layer as the sum of a sparse matrix and a fast transform. Another thing that I found interesting was the notion of matching vs extreme sparsity (<a href="https://openreview.net/forum?id=Ig-VyQc-MLK">Frankle et al., 2021</a>): the former is the sparsity setting where pruned models perform comparatively to the full model while in the latter setting, the performance of pruned models deteriorates. <strong>Strong pruned models should aim to strike a Pareto-optimal sparsity&#8211;performance trade-off.</strong></p><p><strong><a href="http://proceedings.mlr.press/v139/d-ascoli21a/d-ascoli21a.pdf">ConViT: Improving Vision Transformers with Soft Convolutional Inductive Biases</a></strong> This paper is a nice example of how different inductive biases can be combined to make a model more expressive depending on how much data is available. Vision Transformers using self-attention have outperformed CNNs when trained on large datasets as self-attention is more expressive. But CNNs still perform better when trained on smaller datasets due to their inductive bias. The authors propose a slightly modified self-attention layer, which is initialized to act as a convolutional layer. <strong>This way, the model retains the useful inductive bias early during training and can become more expressive if necessary later on.</strong> Such a soft inductive bias may also be of interest for efficient language Transformers, which could limit the attention span early in training (<a href="https://aclanthology.org/P19-1032.pdf">Sukhbaatar et al., 2019</a>).</p><p><strong><a href="http://proceedings.mlr.press/v139/zhao21c/zhao21c.pdf">Calibrate Before Use: Improving Few-Shot Performance of Language Models</a></strong> This paper highlights the instability of prompt-based learning. In particular, prompt-based models are sensitive to the format of the prompt, training examples, and order of examples. A key problem is that the model favours certain answers over others, e.g. answers that are frequent in the prompt, appear towards the end of the prompt, or are frequent in its pre-training data. To address this, the authors propose to first estimate the model's bias using a content-free prompt. The model's predictions can then be recalibrated so that the class scores for the content-free prompt are uniform. <strong>While calibration doesn't alleviate the need for prompt engineering, it reduces the variance when dealing with different prompts.</strong> Given the current popularity of prompt-based methods, this may make working with prompt-based models a lot easier.</p><p><strong><a href="http://proceedings.mlr.press/v139/davis21a/davis21a.pdf">Catformer: Designing Stable Transformers via Sensitivity Analysis</a></strong> This paper introduces the concept of the sensitivity of an architecture, which measures how an architecture's output varies when its parameters are randomly perturbed. The authors also relate this measure to how difficult to train certain architectures are. They then propose a simple modification to the Transformer, which replaces residual connections with concatenation and is more stable to train on a set of reinforcement learning tasks. So far, training difficulty was often discussed anecdotally. <strong>A more principled measure of training difficulty such as sensitivity is a step towards designing not only models that are more powerful but that are also easier to use in real-world applications.</strong></p><p><strong><a href="http://proceedings.mlr.press/v139/koh21a/koh21a.pdf">WILDS: A Benchmark of in-the-Wild Distribution Shifts</a></strong> This is a very diverse benchmark to test how well ML methods generalize across distribution shifts on a wide variety of domains and data modalities. It covers domains as diverse as camera trap photos, cell images, molecular graphs, online comments, and code. <strong>If you are working on robust, modality-agnostic ML methods, then this is the dataset to evaluate on.</strong></p><h2>Open collaboration &#129309;</h2><p>At ICML, I attended the <a href="https://mlcollective.org/icml-2021-open-collab-social/">Social on Open Collaboration in ML Research</a>, hosted by <a href="https://mlcollective.org/">ML Collective</a>, among others. During the event, people shared a diverse range of external collaboration experiences, many of them relating to work done as part of independent research collectives.</p><p>Connor Leahy talked about <a href="https://www.eleuther.ai/about/">EleutherAI</a>, a grassroots collective of researchers who developed not only <a href="https://github.com/EleutherAI/gpt-neo">GPT-Neo</a>, an open-source LM in the style of GPT-3 but also worked on BioML research and ML-generated art (read more about the art below)&#8212;all of this in the past year. <a href="https://blog.eleuther.ai/year-one/">This blog post</a> provides a great overview of their progress so far. To join or contribute, you can head over to their <a href="https://www.eleuther.ai/get-involved/">Discord</a>.</p><p>Edward Elson Kosasih talked about his research as part of <a href="https://mlcollective.org/">ML Collective</a> (MLC), a nonprofit organization dedicated to making ML research accessible. He led a team that <a href="https://arxiv.org/abs/2106.15529">worked on graph neural networks</a> as part of the <a href="https://ogb.stanford.edu/kddcup2021/">Open Graph Benchmark Large Scale Challenge</a>. In order to get involved with MLC, you can <a href="https://mlcollective.org/community/">join their Discord</a>.</p><p>Matthias Gall&#233; discussed the <a href="https://bigscience.huggingface.co/en/#!index.md">BigScience project</a>, also known as The Summer of Language Models 21, a one-year long research workshop on very large language models. The project aims to create and share a large multilingual dataset and to train a very large language model. A diverse set of working groups are dedicated to different parts of the data and model creation process, from data sourcing to prompt engineering, dealing with metadata, and retrieval. To get up to speed on the progress so far, you can watch updates from the first event from July 30, 2021 <a href="https://bigscience.huggingface.co/en/#!pages/events.md">here</a>. To join the project, fill out the form <a href="https://bigscience.huggingface.co/en/#!pages/contact.md">here</a>.</p><p>Salomon Kabongo talked about the work of <a href="https://www.masakhane.io/">Masakhane</a>. Masakhane is a grassroots organisation that aims to strengthen NLP research in African languages. So far, they have released models and datasets for diverse tasks such as machine translation, named entity recognition, and others in many African languages. To get involved, <a href="https://www.masakhane.io/contact">join the Google group and Slack channel</a>.</p><p>On the whole, my impression is that ML and NLP have become much more accessible, in part thanks to research collaborations such as the above, which are open to anyone as long as you're excited and motivated to contribute. Other collaboration opportunities are the <a href="https://forums.fast.ai/">fast.ai</a> or the <a href="https://discuss.huggingface.co/">HuggingFace</a> communities. If you are looking to work in ML or NLP and need collaborators and guidance, I encourage you to join one of the above collaborations.</p><p>For conducting academic collaborations, I shared some lessons of my first external collaboration (and first long paper during my PhD) with Barbara Plank (see below).</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359 424w, https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359 848w, https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359 1272w, https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359 424w, https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359 848w, https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359 1272w, https://s3.amazonaws.com/revue/items/images/010/341/874/original/open_collab.png?1627838359 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">A rough timeline and takeaways of my first external collaboration</figcaption></figure></div><h2>CLIP art &#127912;</h2><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/010/343/521/original/how_clip_generates_art.png?1627847043" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/010/343/521/original/how_clip_generates_art.png?1627847043 424w, https://s3.amazonaws.com/revue/items/images/010/343/521/original/how_clip_generates_art.png?1627847043 848w, https://s3.amazonaws.com/revue/items/images/010/343/521/original/how_clip_generates_art.png?1627847043 1272w, https://s3.amazonaws.com/revue/items/images/010/343/521/original/how_clip_generates_art.png?1627847043 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/010/343/521/original/how_clip_generates_art.png?1627847043" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/010/343/521/original/how_clip_generates_art.png?1627847043&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/010/343/521/original/how_clip_generates_art.png?1627847043 424w, https://s3.amazonaws.com/revue/items/images/010/343/521/original/how_clip_generates_art.png?1627847043 848w, https://s3.amazonaws.com/revue/items/images/010/343/521/original/how_clip_generates_art.png?1627847043 1272w, https://s3.amazonaws.com/revue/items/images/010/343/521/original/how_clip_generates_art.png?1627847043 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption"><a href="https://ml.berkeley.edu/blog/posts/clip-art/?utm_campaign=NLP%20News&amp;utm_medium=email&amp;utm_source=Revue%20newsletter">How CLIP Generates Art (Credit: Charlie Snell)</a></figcaption></figure></div><p>CLIP art, not to be confused with the sometimes slightly cheesy <a href="https://en.wikipedia.org/wiki/Clip_art">type of graphic art</a> often used for illustration purposes, relates to art produced using the <a href="https://openai.com/blog/clip/">CLIP model</a> by OpenAI. CLIP was trained with a contrastive objective to match text with corresponding images. As a result, CLIP is very good at judging which caption best reflects an image, which can be used for zero-shot classification on ImageNet. Alternatively, the model also can be used to gauge which image best suits a description. This is how CLIP is used to generate art (see above), by steering the output of a separate generative model through back-propagation until the model generates an image that matches the description as closely as possible, according to CLIP.</p><p><a href="https://ml.berkeley.edu/blog/posts/clip-art/">This article</a> by Charlie Snell does a great job of charting the development of the art scene that has evolved around this method&#8212;and the often dreamy, impressionistic or psychedelic images that it has produced. The cool thing is that CLIP works with any generative model so the possibilities the method presents develop and grow more diverse as generative models become more powerful. One of my favourite images is the one below of a "a beautiful epic wondrous fantasy painting of the ocean"&nbsp;generated by <a href="https://twitter.com/RiversHaveWings/status/1410020043178446848">@RiversHaveWings</a> using <a href="https://arxiv.org/abs/2106.14843">CLIPDraw</a> + CLIP.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/010/343/690/original/E5Fls_lUYAEEMKW.jpeg?1627847837" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/010/343/690/original/E5Fls_lUYAEEMKW.jpeg?1627847837 424w, https://s3.amazonaws.com/revue/items/images/010/343/690/original/E5Fls_lUYAEEMKW.jpeg?1627847837 848w, https://s3.amazonaws.com/revue/items/images/010/343/690/original/E5Fls_lUYAEEMKW.jpeg?1627847837 1272w, https://s3.amazonaws.com/revue/items/images/010/343/690/original/E5Fls_lUYAEEMKW.jpeg?1627847837 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/010/343/690/original/E5Fls_lUYAEEMKW.jpeg?1627847837" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/010/343/690/original/E5Fls_lUYAEEMKW.jpeg?1627847837&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/010/343/690/original/E5Fls_lUYAEEMKW.jpeg?1627847837 424w, https://s3.amazonaws.com/revue/items/images/010/343/690/original/E5Fls_lUYAEEMKW.jpeg?1627847837 848w, https://s3.amazonaws.com/revue/items/images/010/343/690/original/E5Fls_lUYAEEMKW.jpeg?1627847837 1272w, https://s3.amazonaws.com/revue/items/images/010/343/690/original/E5Fls_lUYAEEMKW.jpeg?1627847837 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption"><a href="https://twitter.com/RiversHaveWings/status/1410020043178446848?ref_src=twsrc%5Etfw%7Ctwcamp%5Etweetembed%7Ctwterm%5E1410020043178446848%7Ctwgr%5E%7Ctwcon%5Es1_&amp;ref_url=https%3A%2F%2Fdisqus.com%2Fembed%2Fcomments%2F%3Fbase%3Ddefaultf%3Dmlab-blogt_u%3Dhttps3A2F2Fml.berkeley.edu2Fblog2Fposts2Fclip-art2Ft_e%3D5Bobject20Object5Dt_d%3DAlien20Dreams3A20An20Emerging20Art20Scenet_t%3D5Bobject20Object5Ds_o%3Ddefaultversion%3D7302391be467f75d298eac65b5cfa2cc&amp;utm_campaign=NLP%20News&amp;utm_medium=email&amp;utm_source=Revue%20newsletter">"a beautiful epic wondrous fantasy painting of the ocean" by @RiversHaveWings</a></figcaption></figure></div><h2>Internet augmentation &#128187;</h2><p>Current large language models are trained on large amounts of unlabelled data, mainly from the web. However, they do not yet leverage everything the Internet has to offer. In other words, there are many types of signals on the web that are currently not used for learning. Barbara Plank has called this <a href="https://bplank.github.io/#FortuitousData">fortuitous data</a>, as such data is often available by accident or good fortune.</p><p>A great example of such fortuitous data is the HTML structure underlying web pages. Such structure can provide both useful learning signal for a model as well as be useful for generating prompts, by letting the model auto-complete the HTML structure of a document. <a href="https://arxiv.org/abs/2107.06955">Aghajanyan et al. (2021)</a> recently proposed the cleverly named HTLM, a large language model trained on HTML structure. They show that the model excels at zero-shot natural language generation using structured HTML-based prompts. In addition, they propose to control the size of the generated output sequence by using size hints, noisy estimates of the length of the generated span inserted right after the MASK token.</p><p>Another recent example of leveraging more of what the web has to offer is an extension of <a href="https://ruder.io/research-highlights-2020/index.html#2-retrieval-augmentation">retrieval augmentation</a> to the Internet. Specifically, rather than learning to retrieve relevant information only from a large corpus of unlabelled text, a model can learn to retrieve from the entire Internet. To make this feasible, <a href="https://arxiv.org/abs/2107.07566">Komeili et al. (2021)</a> learn to generate a search query based on the context of a dialogue. They then condition on the search results to generate a response. The resulting Internet-augmented dialogue model outperforms both the use of retrieval augmentation and no augmentation.</p><p>Other forms of information that so far have been neglected are a) the information from hyperlinked pages, which could be used for conditioning during training; b) hyperlink patterns, to learn which information to trust; c) multi-modal context on webpages, to ground representations; d) snapshots of webpages over time, to learn re-writing, etc; e) timestamps for modelling time (<a href="https://arxiv.org/abs/2106.15110">Dhingra et al., 2021</a>), f) content by the same users across multiple websites for authorship and style modelling, etc.</p><h2>New GLUE-style benchmarks &#127963;</h2><p>Since the development of models that learn general representations, mainly via self-supervised learning, it has become common to evaluate such models on benchmarks comprising a diverse set of different tasks. The most prominent of these are arguably the <a href="https://gluebenchmark.com/">GLUE</a> and <a href="https://super.gluebenchmark.com/">SuperGLUE</a> benchmarks. Following in their footsteps, benchmarks serving a wide array of settings and languages have been proposed.</p><p>I'm particularly excited about two recent additions to this ever-growing evaluation environment. <strong>F</strong>ew-shot <strong>L</strong>anguage <strong>E</strong>valuation across (<strong>X</strong>) many transfer types (FLEX; <a href="https://arxiv.org/abs/2107.07170">Bragg and Cohan et al., 2021</a>) is a benchmark focused on few-shot NLP, something I hoped to <a href="https://ruder.io/requests-for-research/index.html#fewshotlearningfornlp">encourage in 2018</a>. It not only covers the standard meta-learning/few-shot learning setup with separate meta-training and meta-test portions but also manages to capture the current zeitgeist by evaluating zero-shot evaluation based on textual descriptions.</p><p>The second benchmark is the <strong>S</strong>peech processing <strong>U</strong>niversal <strong>Per</strong>formance <strong>B</strong>enchmark (SUPERB; <a href="https://arxiv.org/abs/2105.01051">Yang et al., 2021</a>), which aims to do for speech what GLUE has done for NLP, by providing a general platform to evaluate self-supervised speech models on 10 different tasks. The benchmark covers core speech tasks, from modelling content (phonemes, transcription, keywords) and speakers (identification, verification, and diarization) to dealing with semantics (intents, slot filling) and paralinguistic features (emotions). Such a standardization will likely open the door to the development of more powerful self-supervised speech models.</p><p>Another type of benchmark I'm excited about is one that covers many tasks in a given language. I've recently had the chance to contribute to two such benchmarks: <a href="https://openreview.net/forum?id=JH61CD7afTv">LiRo</a> for NLU tasks in Romanian and <a href="https://arxiv.org/abs/2104.08200">IndoNLG</a> for NLG tasks in Indonesian. Facilitating evaluation on a diverse set of tasks in a given language is one of the best ways to incentivise progress in that language in my opinion.</p>]]></content:encoded></item><item><title><![CDATA[GitHub Copilot, The Perceiver, Beyond the Transformer, Data augmentation, NL augmenter 🦎 → 🐍, Research communication]]></title><description><![CDATA[Hi all,This newsletter is a bit delayed. I had to skip the last one as I had to take a break after a busy period in the end of May (EMNLP and NeurIPS deadlines). At the same time, there's so much happening that I've found it hard to catch up. Now I'm back, feeling more energized, and updating myself (and you) on what's new.I'll discuss the biggest advances over the last months including GitHub Copilot, the Perceiver, and non-self-attention models.I'll also talk about something that is challenging for me when writing this newsletter: striking the right balance between content that is both timely but also relevant in the long-term. TL;DR: I'm planning to keep newsletters somewhat shorter in the future to have more time for in-depth blog posts.I really appreciate your feedback, so let me know what you&#160;love &#10084;&#65039; and hate &#128148; about this edition. Simply hit reply on the issue.Click here to view this newsletter in your browser.If you were referred by a friend, click&#160;here&#160;to subscribe. If you enjoyed this issue,&#160;give it a&#160;tweet&#160;&#128037;.]]></description><link>https://newsletter.ruder.io/p/github-copilot-the-perceiver-beyond-the-transformer-data-augmentation-nl-augmenter-research-communication-527358</link><guid isPermaLink="false">https://newsletter.ruder.io/p/github-copilot-the-perceiver-beyond-the-transformer-data-augmentation-nl-augmenter-research-communication-527358</guid><dc:creator><![CDATA[Sebastian Ruder]]></dc:creator><pubDate>Mon, 19 Jul 2021 09:00:02 GMT</pubDate><enclosure url="https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi all,</p><p>This newsletter is a bit delayed. I had to skip the last one as I had to take a break after a busy period in the end of May (EMNLP and NeurIPS deadlines). At the same time, there's so much happening that I've found it hard to catch up. Now I'm back, feeling more energized, and updating myself (and you) on what's new.</p><p>I'll discuss the biggest advances over the last months including GitHub Copilot, the Perceiver, and non-self-attention models.</p><p>I'll also talk about something that is challenging for me when writing this newsletter: striking the right balance between content that is both timely but also relevant in the long-term. <strong>TL;DR:</strong> I'm planning to keep newsletters somewhat shorter in the future to have more time for in-depth blog posts.</p><p>I really appreciate your feedback, so let me know what you&nbsp;love &#10084;&#65039; and hate &#128148; about this edition. Simply hit reply on the issue.</p><p><em>Click <strong><a href="https://newsletter.ruder.io/issues/github-copilot-the-perceiver-beyond-the-transformer-data-augmentation-nl-augmenter-research-communication-527358/a35ce65f-a2ee-4ba6-9316-e617f788fbf7">here</a></strong> to view this newsletter in your browser.</em></p><p>If you were referred by a friend, click&nbsp;<a href="http://newsletter.ruder.io/?utm_campaign=NLP%20News&amp;utm_medium=email&amp;utm_source=Revue%20newsletter">here</a>&nbsp;to subscribe. If you enjoyed this issue,&nbsp;give it a&nbsp;<a href="https://twitter.com/intent/tweet?url=http%3A%2F%2Fnewsletter.ruder.io%2Fissues%2Fgithub-copilot-the-perceiver-beyond-the-transformer-data-augmentation-nl-augmenter-research-communication-527358%2Fa35ce65f-a2ee-4ba6-9316-e617f788fbf7&amp;via=revue&amp;text=GitHub%20Copilot%2C%20The%20Perceiver%2C%20Beyond%20the%20Transformer%2C%20Data%20augmentation%2C%20NL%20augmenter%20%F0%9F%A6%8E%20%E2%86%92%20%F0%9F%90%8D%2C%20Research%20communication%20by%20%40seb_ruder&amp;related=revue">tweet</a>&nbsp;&#128037;.</p><div><hr></div><h2>May&#8211;July round-up</h2><h2>OpenAI Codex / GitHub Copilot</h2><p>If you are working with software, then you've probably heard about the release of <a href="https://copilot.github.com/">GitHub Copilot</a>, a coding assistant based on Codex, a GPT language model fine-tuned on code on GitHub (see <a href="https://arxiv.org/abs/2107.03374">the paper</a>). As far as I'm aware, this represents one of the first products of a large company where a user directly interacts with a large generative language model. Large language models are also used in many other applications such as <a href="https://blog.google/products/search/search-language-understanding-bert/">Google Search</a>, but such applications typically include a wide array of other signals.</p><p>There are a couple of interesting take-aways from the <a href="https://arxiv.org/abs/2107.03374">paper</a>: One is that the model was not trained from scratch but an existing GPT model (up to 12B parameters, so not the largest GPT-3 model; the deployed model may be larger, however) was <a href="https://ruder.io/recent-advances-lm-fine-tuning/#adaptive-fine-tuning">adaptively fine-tuned</a> on code from GitHub. In addition, they <a href="https://ruder.io/recent-advances-lm-fine-tuning/#behavioural-fine-tuning">further fine-tune</a> two variants of the model (Codex-S and Codex-G) to generate stand-alone functions and docstrings respectively.</p><p>Given that language models are prone to reproduce inputs (<a href="https://arxiv.org/abs/2012.07805">Raffel et al., 2020</a>) people have already found memorization issues with Copilot such as <a href="https://mobile.twitter.com/kylpeacock/status/1410749018183933952?utm_campaign=NLP%20News&amp;utm_medium=email&amp;utm_source=Revue%20newsletter">copy-pasting a person's contact info</a>. One issue is that the Codex model was trained on all code on GitHub including code with potentially problematic licenses. <a href="https://twitter.com/alexjc/status/1410524416660889607?s=20">GitHub</a> <a href="https://twitter.com/natfriedman/status/1409914420579344385?s=20">claims</a> that code produced by an AI model is "fair use"&#8212;it's controversial whether this is actually the case, given that the model may reproduce passages verbatim.</p><p>Another question is whether Copilot will be able to make a meaningful difference in the workflow of programmers. In a <a href="https://newsletter.ruder.io/issues/eacl-iclr-naacl-papers-round-up-research-reality-checks-ml-on-code-592784">previous newsletter</a>, I discussed current ML on code work. In particular, I highlighted <a href="https://arxiv.org/abs/2101.11149">a study</a>, which found that current models did not improve productivity or code quality when used for in-IDE code generation. A practical limitation of Copilot is that it only considers the code in the current file (rather than in the entire codebase) and can thus only generate relatively self-contained code. So it remains to be seen whether it will provide meaningful benefits beyond the capabilities of existing models.</p><p>GitHub markets Copilot as an "AI pair programmer". <a href="https://en.wikipedia.org/wiki/Pair_programming">Pair programming</a> is essentially a form of dialogue grounded in code. Similar to interacting with a dialogue agent, one requirement for successfully completing a task is a shared foundation of meaning. A task-oriented dialogue agent needs to know about the relevant entities and intents that are necessary for, say, booking a restaurant. In the same vein, an effective pair programming assistant should also have knowledge of the underlying codebase and its functions and variables.</p><p>Similar to how conversational question answering has been a focus in the community of late (see <a href="https://arxiv.org/abs/2106.00874">this recent paper</a> for an overview), a conversational pair programming task would be a great way to measure progress regarding not just whether a model can produce a given function but whether it can effectively collaborate with a human. Given the promise of such models for augmenting the programming workflow, expect to see much more work in this space.</p><h2><strong>The Perceiver</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011 424w, https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011 848w, https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011 1272w, https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011 424w, https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011 848w, https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011 1272w, https://s3.amazonaws.com/revue/items/images/010/136/581/original/perceiver.png?1626454011 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The Perceiver uses cross-attention to project a large input byte array to a small latent array, which is processed with a regular Transformer stack. Cross-attention and Transformer stacks are interleaved throughout the model, with optional parameter sharing.</figcaption></figure></div><p><a href="https://arxiv.org/abs/2103.03206">The Perceiver</a> (Jaegle et al., ICML 2021) is one of the recent models that I'm most excited about. The main motivation of the work is to enable a Transformer-like architecture to scale to very high-dimensional inputs (<a href="https://arxiv.org/abs/2010.11929">Vision Transformers</a> are typically applied to image patches to overcome the computational complexity of self-attention). There have been a lot of recent more efficient Transformer architectures (see <a href="https://arxiv.org/abs/2009.06732">this paper</a> for an overview) but these still depend on the length of the input, typically linearly.</p><p>In contrast, the Perceiver uses a latent array of a fixed dimensionality as its base representation (see above). This representation is then conditioned via cross-attention (as in a standard encoder-decoder model) on the much larger input array and then processed with a Transformer stack, in alternating fashion. If parameters are shared across Transformer blocks and cross-attention layers, the Perceiver can essentially be seen as an RNN with a Transformer at its core. It is also similar in spirit to the <a href="https://arxiv.org/abs/1807.03819">Universal Transformer</a> (Dehghani et al., ICLR 2019), a model that applies the same Transformer block to an input multiple times.</p><p>The authors apply the Perceiver to three datasets across different modalities (ImageNet, video and audio, and 3D point clouds) and report performance competitive with the state of the art on all of them. You can also check out <a href="https://www.youtube.com/watch?v=P_xeshTnPZg">Yannic Kilcher's video</a> for a more visual introduction and contextualisation of the Perceiver.</p><h2><strong>Beyond the Transformer</strong></h2><p>Another recent trend has been the emergence of models that seek to replace the ubiquitous self-attention layer, most notably using <a href="https://en.wikipedia.org/wiki/Multilayer_perceptron">multilayer perceptrons</a> (MLPs). The <a href="https://arxiv.org/abs/2105.01601">MLP-Mixer</a> (Tolstikhin et al., 2021) applies MLPs independently to image patches as well as across patches and achieves competitive results on image classification tasks. <a href="https://arxiv.org/abs/2105.08050">Liu et al. (2021)</a> propose gMLP, a gated MLP architecture that achieves performance similar to Transformers on NLP and vision tasks.</p><p>A non-MLP based recent model is <a href="https://arxiv.org/abs/2105.03824">FNet</a> (Lee-Thorp, 2021), which uses 1D Fourier Transforms instead of self-attention to mix information at the token level. While the model is less expressive than self-attention based models such as BERT, it is much faster and still achieves competitive results in many settings.</p><p>Another thread of work in this area revisits the dominance of self-attention by applying the same treatment to convolutions (<a href="https://arxiv.org/abs/2105.03322">Tay et al., ACL 2021</a>): It turns out that if CNNs are pre-trained the same way as Transformer models, they achieve competitive performance on many NLP tasks. They mainly underperform on tasks that require modelling relations across sentences (such as paraphrasing, NLI, or QA), tasks that are notably over-represented on standard benchmarks such as GLUE.</p><p>On a similar note, a recent paper (<a href="https://arxiv.org/abs/2107.07002">Dehghani et al., 2021</a>) by some of the same authors argues that the tasks we focus on as part of a benchmark induce a bias in terms of the models that will succeed. If standard benchmarks such as GLUE were constructed differently, would we still have ended up with self-attention-based models dominating or would CNN-based models be much more common?</p><p>In sum, an MLP may unfortunately not be all you need. However, while the hegemony of self-attention may still endure, recent challengers based on MLPs, convolutions, and various other transformations encourage us to rethink the fundamental building blocks of our models.</p><h2>Data augmentation, NL augmenter &#129422; &#8594; &#128013;</h2><p>Data augmentation is a common tool used in computer vision but much less common in NLP. NLP is more challenging for augmentation due to the discrete nature of language, which also makes it harder to preserve meaning across transformations. This recent survey (<a href="https://arxiv.org/abs/2105.03075">Feng et al., ACL Findings 2021</a>) gives an overview of recent approaches in this area. In particular, one thing that is missing for current data augmentation approaches in NLP is a unified benchmark and framework where many different approaches can be tried and compared to each other.</p><p>Towards this goal, <a href="https://gem-benchmark.com/nl_augmenter">NL-Augmenter</a> is a collaborative effort that aims to collect a wide range of transformations, perturbations, and filters that generate additional data either for training or to test model robustness.</p><p>It is motivated by recent efforts such as the Beyond the Imitation Game Benchmark (<a href="https://github.com/google/BIG-bench">BIG-bench</a>), a collaborative project that crowd-sourced tasks to probe large language models. BIG-bench project has attracted a large amount of interest with people <a href="https://github.com/google/BIG-bench/pulls?q=is%3Apr">proposing 100s of tasks</a>.</p><p>NL-Augmenter invites submissions <a href="https://github.com/GEM-benchmark/NL-Augmenter">via GitHub pull requests</a>. Submitted transformations may augment data in diverse ways such as introducing spelling errors, translating to a different language, randomizing names and numbers, paraphrasing, etc. Some of my favourite submissions introduce transformations that <a href="https://github.com/GEM-benchmark/NL-Augmenter/pull/100">randomly swap words that sound similar to each other</a>, <a href="https://github.com/GEM-benchmark/NL-Augmenter/pull/84">replace names with more gender and culturally diverse ones</a>, or <a href="https://github.com/GEM-benchmark/NL-Augmenter/pull/40">translate random words to another language</a>.</p><p>The submission deadline is September 1, 2021. If you are interested in data augmentation for NLP, this is a great chance to contribute to a large community effort.</p><h2>The research communication continuum &#128483;</h2><p>I think a lot about how to communicate research effectively. One thing that is on my mind lately is what I call&#8212;for lack of a better word&#8212;the 'research communication continuum', essentially <strong>to what extent</strong> and <strong>in what form</strong> to discuss a given research topic. I have depicted different levels of content and the most common formats below in order of increasing complexity, but of course different formats are possible for each.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/009/947/160/original/research_communication_continuum.png?1625399201" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/009/947/160/original/research_communication_continuum.png?1625399201 424w, https://s3.amazonaws.com/revue/items/images/009/947/160/original/research_communication_continuum.png?1625399201 848w, https://s3.amazonaws.com/revue/items/images/009/947/160/original/research_communication_continuum.png?1625399201 1272w, https://s3.amazonaws.com/revue/items/images/009/947/160/original/research_communication_continuum.png?1625399201 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/009/947/160/original/research_communication_continuum.png?1625399201" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/009/947/160/original/research_communication_continuum.png?1625399201&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/009/947/160/original/research_communication_continuum.png?1625399201 424w, https://s3.amazonaws.com/revue/items/images/009/947/160/original/research_communication_continuum.png?1625399201 848w, https://s3.amazonaws.com/revue/items/images/009/947/160/original/research_communication_continuum.png?1625399201 1272w, https://s3.amazonaws.com/revue/items/images/009/947/160/original/research_communication_continuum.png?1625399201 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">The research communication continuum</figcaption></figure></div><p>There are a lot of great sources to get highlights or digests of interesting new articles such as SkyNet Today's <a href="https://lastweekin.ai/">Last Week in AI</a>, Yannic Kilcher's <a href="https://www.youtube.com/c/YannicKilcher/videos">ML News</a>, <a href="https://read.deeplearning.ai/the-batch/">The Batch</a>, <a href="https://jack-clark.net/">Import AI</a>, and many more.</p><p>To complement these resources, I have tried to focus more on in-depth discussions with this newsletter. However, such content is more time-consuming to write and takes time away from writing more comprehensive blog posts.</p><p>Going forward, I'm planning to focus more on slightly lighter, opinionated takes on current research themes rather than super in-depth discussions in this newsletter. That should give me more time go deep in blog posts and to explore exciting research topics with you more regularly. Stay tuned!</p><h2>Rethinking ML Papers &#128221;</h2><p>Talking about research communication, the <a href="https://rethinkingmlpapers.github.io/">Rethinking ML Papers</a> workshop at ICLR 2021 explored just this topic and featured many luminaries of the ML communication space (if you registered at ICLR, you can view the content of the workshop <a href="https://iclr.cc/virtual/2021/workshop/2142">here</a>). My highlights were:</p><ul><li><p>Lilian Weng who <a href="https://lilianweng.github.io/lil-log/">writes excellent overview articles</a> about a diverse set of topics, from contrastive representation learning to controllable text generation. She talked about how you can <a href="https://twitter.com/rethinkmlpapers/status/1390700606772371458">catch up with the field by writing a high-quality ML blog post</a> (also one of my motivations for writing blog posts).</p></li><li><p>Terence Parr who writes <a href="https://explained.ai/decision-tree-viz/index.html">visual, lucid articles</a> explaining core ML concepts. His post on <a href="https://explained.ai/decision-tree-viz/index.html">visualizing decision trees</a> is still one of the most visually intuitive illustrations of the method that I have seen. He spoke about the <a href="https://twitter.com/rethinkmlpapers/status/1390707375481491466?s=20">role of visualization in ML</a>.</p></li><li><p>David Ha, <a href="https://otoro.net/ml/">interactive visualization virtuoso</a>, talked about <a href="https://twitter.com/rethinkmlpapers/status/1390654809649332227?s=20">interactive web demos of ML models</a>.</p></li><li><p>Jay Alammar, of <a href="https://jalammar.github.io/illustrated-transformer/">Illustrated Transformer</a> fame, discussed <a href="https://twitter.com/JayAlammar/status/1386952976925958145?s=20">how to communicate ML research via illustrated and interactive web articles</a>.</p></li></ul><h2>Fun papers</h2><blockquote><p><em>And Now For Something Completely Different...</em></p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/010/136/542/original/1-s2.0-S0168159121001258-gr4.jpg?1626453845" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/010/136/542/original/1-s2.0-S0168159121001258-gr4.jpg?1626453845 424w, https://s3.amazonaws.com/revue/items/images/010/136/542/original/1-s2.0-S0168159121001258-gr4.jpg?1626453845 848w, https://s3.amazonaws.com/revue/items/images/010/136/542/original/1-s2.0-S0168159121001258-gr4.jpg?1626453845 1272w, https://s3.amazonaws.com/revue/items/images/010/136/542/original/1-s2.0-S0168159121001258-gr4.jpg?1626453845 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/010/136/542/original/1-s2.0-S0168159121001258-gr4.jpg?1626453845" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/010/136/542/original/1-s2.0-S0168159121001258-gr4.jpg?1626453845&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/010/136/542/original/1-s2.0-S0168159121001258-gr4.jpg?1626453845 424w, https://s3.amazonaws.com/revue/items/images/010/136/542/original/1-s2.0-S0168159121001258-gr4.jpg?1626453845 848w, https://s3.amazonaws.com/revue/items/images/010/136/542/original/1-s2.0-S0168159121001258-gr4.jpg?1626453845 1272w, https://s3.amazonaws.com/revue/items/images/010/136/542/original/1-s2.0-S0168159121001258-gr4.jpg?1626453845 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">If I fits, I sits.</figcaption></figure></div><p><strong><a href="https://www.sciencedirect.com/science/article/pii/S0168159121001258">If I fits I sits: A citizen science investigation into illusory contour susceptibility in domestic cats (Felis silvestris catus)</a></strong> <strong>(Applied Animal Behaviour Science, 2021)</strong> From a different subject area, this article is a large-scale study that capitalizes on two important trends: 1) citizen science, which emphasizes public participation and collaboration in research, and 2) the Internet's love of cats. It turns out that cats not only prefer to sit in physical box-like spaces but also tend to do so if enclosures are illusory, such as the <a href="https://en.wikipedia.org/wiki/Illusory_contours">Kanizsa square</a> visual illusion.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://s3.amazonaws.com/revue/items/images/010/153/948/original/elmo_costume.png?1626610324" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://s3.amazonaws.com/revue/items/images/010/153/948/original/elmo_costume.png?1626610324 424w, https://s3.amazonaws.com/revue/items/images/010/153/948/original/elmo_costume.png?1626610324 848w, https://s3.amazonaws.com/revue/items/images/010/153/948/original/elmo_costume.png?1626610324 1272w, https://s3.amazonaws.com/revue/items/images/010/153/948/original/elmo_costume.png?1626610324 1456w" sizes="100vw"><img src="https://s3.amazonaws.com/revue/items/images/010/153/948/original/elmo_costume.png?1626610324" data-attrs="{&quot;src&quot;:&quot;https://s3.amazonaws.com/revue/items/images/010/153/948/original/elmo_costume.png?1626610324&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://s3.amazonaws.com/revue/items/images/010/153/948/original/elmo_costume.png?1626610324 424w, https://s3.amazonaws.com/revue/items/images/010/153/948/original/elmo_costume.png?1626610324 848w, https://s3.amazonaws.com/revue/items/images/010/153/948/original/elmo_costume.png?1626610324 1272w, https://s3.amazonaws.com/revue/items/images/010/153/948/original/elmo_costume.png?1626610324 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Lecturers going the extra mile for their students by dressing up in NLP-themed costumes (here: ELMo)</figcaption></figure></div><p><strong><a href="https://aclanthology.org/2021.teachingnlp-1.2/">Teaching a Massive Open Online Course on Natural Language Processing (Teaching NLP Workshop 2021)</a></strong> Teaching a new course, particularly during the COVID pandemic can be incredibly challenging. This is a nice example of a massive open online NLP course that lecturers from Moscow taught. In the paper, they share their 12-week syllabus, consisting of lectures covering both fundamentals as well as recent work, real-time coding sessions, and interviews with experts. One thing that surely contributed to the course's success are the thematically fitting wardrobe choices of the lecturers who dressed in Sesame street kigurumis (see above).</p>]]></content:encoded></item></channel></rss>