The study confirms that RECKONING's ability to disentangle relevant knowledge is maintained even when scaling the model size using GPT-2-XL-LoRAThe study confirms that RECKONING's ability to disentangle relevant knowledge is maintained even when scaling the model size using GPT-2-XL-LoRA

Distractor Robustness: RECKONING Significantly Outperforms FT-ICR in Reasoning Over Irrelevant Facts

Abstract and 1. Introduction

  1. Background

  2. Method

  3. Experiments

    4.1 Multi-hop Reasoning Performance

    4.2 Reasoning with Distractors

    4.3 Generalization to Real-World knowledge

    4.4 Run-time Analysis

    4.5 Memorizing Knowledge

  4. Related Work

  5. Conclusion, Acknowledgements, and References

\ A. Dataset

B. In-context Reasoning with Distractors

C. Implementation Details

D. Adaptive Learning Rate

E. Experiments with Large Language Models

4.2 Reasoning with Distractors

In cases where multiple questions must be answered about the same knowledge set, some knowledge that is relevant to one question will likely be irrelevant to another question. For example, in Table 7, the fact “Charlie is White.” is not needed to answer the question “Harry is red?”. Thus, it is important to evaluate the robustness of RECKONING when there exists irrelevant information (i.e., distractors) in the knowledge set. In this experiment, we analyze RECKONING’s ability to focus on the correct knowledge and ignore distractors when answering questions. We use ProofWriter as the evaluation dataset since it already has a setting with distractors included in the knowledge. For systematic analysis, we gradually add distractors to the context (starting from 2 and finishing at all possible distractors, of which there are an average of 7 per question). We train RECKONING and the baseline using the multi-task objective, where the model must (1) recall all of the facts and rules relevant to the question and (2) predict the conclusion based on the correct knowledge. In this case, we adapt training such that for each question x, the outer-loop (Equation (5)) CLM loss is only computed with respect to the relevant facts from K, thereby learning to recall only relevant facts during training.

\ In Figure 5, we see that RECKONING’s performance is consistently more robust under distractors than the FT-ICR baseline. When we include all of the distractors in the context, RECKONING achieves a significantly higher average label accuracy (82.5%) across hops than the baseline (70.9%), as computed by the average of the 3 considered hop depths. Additionally, compared to performance with no distractors, RECKONING’s performance only drops 17.1% while the baseline performance drops 28.6%, thereby exhibiting a better ability to disentangle the correct knowledge from the distractors.

\ Finally, we also explore RECKONING’s generalizability to models with a larger parameter size. We scale up the language model we used, GPT-2-small (124M), to GPT-2-XL (1.5B) by adopting a parameter efficient finetuning method LoRA [33]. For simplicity, we only evaluate the models on the most difficult settings, i.e., ProofWriter-5-hop with all the distractors. With GPT-2-XL-LoRA, in-context reasoning achieves 65% accuracy on the test set, while our RECKONING model achieves 70.2% accuracy, a 5% performance gain. This result suggests that RECKONING’s advantages in the presence of distractors hold even as models scale in size.

\

:::info Authors:

(1) Zeming Chen, EPFL (zeming.chen@epfl.ch);

(2) Gail Weiss, EPFL (antoine.bosselut@epfl.ch);

(3) Eric Mitchell, Stanford University (eric.mitchell@cs.stanford.edu)';

(4) Asli Celikyilmaz, Meta AI Research (aslic@meta.com);

(5) Antoine Bosselut, EPFL (antoine.bosselut@epfl.ch).

:::


:::info This paper is available on arxiv under CC BY 4.0 DEED license.

:::

\

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Trump criticized the unusual phenomenon of "good news not driving prices up" and warned dissidents not to even think about taking the helm of the Federal Reserve.

Trump criticized the unusual phenomenon of "good news not driving prices up" and warned dissidents not to even think about taking the helm of the Federal Reserve.

PANews reported on December 24th that US President Trump praised the third-quarter GDP data on social media, noting that GDP growth reached 4.2%, far exceeding
Share
PANews2025/12/24 08:16
Is Doge Still The Best Crypto Investment, Or Will Pepeto Make You Rich In 2025

Is Doge Still The Best Crypto Investment, Or Will Pepeto Make You Rich In 2025

The post Is Doge Still The Best Crypto Investment, Or Will Pepeto Make You Rich In 2025 appeared on BitcoinEthereumNews.com. Crypto News 18 September 2025 | 13:39 Is Dogecoin actually running out of gas, after making people millionaires overnight? As investors hunt for the best crypto to buy now and the best crypto to invest in 2025, Dogecoin still owns the meme spotlight, yet its upside looks capped according to today’s Dogecoin price prediction. Focus is shifting toward projects that marry community with real on chain utility. People searching best crypto to buy now want shipped products, audits, and transparent tokenomics. That frames the honest matchup for this cycle, Dogecoin versus Pepeto. Meet Pepeto, an Ethereum based meme coin built with live rails, PepetoSwap for zero fee trading and Pepeto Bridge for smooth cross chain moves. By blending story with tools people can touch today, and speaking directly to crypto presale 2025 demand, Pepeto puts utility, clarity, and distribution first. In a market where older meme coins risk drifting on sentiment, Pepeto’s delivery gives it a credible seat in the best crypto investment debate. First, here is why Dogecoin may be fading. Dogecoin Price Prediction Is Dogecoin Losing Momentum Remember when Dogecoin made crypto feel effortless. In 2013, Doge turned an internet joke into money and a movement that welcomed everyone. A decade later the market is tougher and the relentless tailwind is gone, sentiment is choppier and patience matters. With Doge near $0.268, the setup reads bearish to neutral for the next few weeks. If the $0.26 shelf holds on daily closes, expect choppy range trading toward $0.29 to $0.30 where rallies keep stalling. Lose $0.26 and momentum often slides into $0.245 with risk of a deeper probe toward $0.22 to $0.21. Close back above $0.30 and the downside bias is likely neutralized, opening room for a squeeze into the low $0.30s. Beyond the price view, Dogecoin still centers…
Share
BitcoinEthereumNews2025/09/18 18:56
Russia Proposes Crypto Access for Retail Investors via Knowledge Tests, $3,834 Annual Cap

Russia Proposes Crypto Access for Retail Investors via Knowledge Tests, $3,834 Annual Cap

Russia's central bank has submitted a draft proposal that would permit non-qualified investors to purchase certain cryptocurrencies after passing a mandatory knowledge test, with annual purchases capped at approximately $3,834. The proposal represents a significant shift in Russia's approach to cryptocurrency regulation, balancing controlled retail access with investor protection measures amid the country's evolving digital asset policy.
Share
MEXC NEWS2025/12/24 10:27