Robuta

https://openreview.net/forum?id=Le9anH3kv1&referrer=%5Bthe%20profile%20of%20Sam%20Havens%5D(%2Fprofile%3Fid%3D~Sam_Havens1)
Retrieval Augmented Generation (RAG) has emerged as a crucial technique for enhancing the accuracy of Large Language Models (LLMs) by incorporating external...
large language modelslong contextragperformanceopenreview
https://aclanthology.org/2022.naacl-main.144/
Wenhan Xiong, Barlas Oguz, Anchit Gupta, Xilun Chen, Diana Liskovich, Omer Levy, Scott Yih, Yashar Mehdad. Proceedings of the 2022 Conference of the North...
long contextsimplelocalremaincompetitive
https://smollm3.org/
SmolLM3 is a powerful 3B parameter language model designed for efficient reasoning, long context understanding, and multilingual applications. Explore the...
long contextsmolmultilingualreasoner
https://arxiv.org/abs/2406.10149
Abstract page for arXiv paper 2406.10149: BABILong: Testing the Limits of LLMs with Long Context Reasoning-in-a-Haystack
testinglimitsllmslong
https://pubmed.ncbi.nlm.nih.gov/19675288/?dopt=Abstract
Acoustic processing requires integration over time. We have used in vivo intracellular recording to measure neuronal integration times in anesthetized rats....
long lastingneural encodingcontextdependencemodels
https://openreview.net/forum?id=ulCAPXYXfa&referrer=%5Bthe%20profile%20of%20Bo%20Zheng%5D(%2Fprofile%3Fid%3D~Bo_Zheng8)
During the inference phase of Large Language Models (LLMs) with long context, a substantial portion of GPU memory is allocated to the KV cache, with memory...
dynamiccontextselectionefficientlong
https://openreview.net/forum?id=ulCAPXYXfa&referrer=%5Bthe%20profile%20of%20Tian%20Wang%5D(%2Fprofile%3Fid%3D~Tian_Wang13)
During the inference phase of Large Language Models (LLMs) with long context, a substantial portion of GPU memory is allocated to the KV cache, with memory...
dynamiccontextselectionefficientlong
https://the-decoder.com/mistrals-new-code-model-features-a-32k-context-window-for-unmatched-long-code-support/
French AI company Mistral has launched Codestral, a new coding model that delivers high coding performance with less computational overhead than existing...
new codecontext windowmistralmodelfeatures
https://openreview.net/forum?id=2QG6QHUehV&referrer=%5Bthe%20profile%20of%20Kaikai%20An%5D(%2Fprofile%3Fid%3D~Kaikai_An1)
Aligning large language models to handle instructions with extremely long contexts has yet to be fully investigated. Previous studies attempt to scale up the...
long contextgateauselectinginfluentialsamples
https://openreview.net/forum?id=Eh0Od2BJIM&referrer=%5Bthe%20profile%20of%20Amir%20Zandieh%5D(%2Fprofile%3Fid%3D~Amir_Zandieh2)
We present an approximate attention mechanism named `HyperAttention` to address the computational challenges posed by the growing complexity of long contexts...
long contextlinear timeattentionopenreview
https://aclanthology.org/2024.emnlp-demo.21/
Gefei Gu, Yilun Zhao, Ruoxi Ning, Yanan Zheng, Arman Cohan. Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: System...
long contexttailtoolkitautomaticrealistic
https://www.mdpi.com/2071-1050/15/21/15674
The rapid ageing of populations around the world is creating complex challenges for national governments.
elderly carelong termdependencycost
https://www.context-pack.com/
Move AI memory across platforms with Context Pack. Create portable memory packs from ChatGPT, Claude, and Gemini chats. Transfer your AI conversation history...
long term memorycontextpackai
https://the-decoder.com/metas-llama-4-models-show-promise-on-standard-tests-but-struggle-with-long-context-tasks/
Apr 12, 2025 - New independent evaluations reveal that Meta's latest Llama 4 models - Maverick and Scout - perform well in standard tests but struggle with complex...
metallamamodelsshowpromise
https://www.databricks.com/blog/long-context-rag-capabilities-openai-o1-and-google-gemini?utm_source=chatgpt.com
long contextragcapabilitiesopenaigoogle
https://scale.com/blog/long-context-instruction-following
A Guide to Improving Long Context Instruction Following on Open Source Models
long contextguideimprovinginstructionfollowing
https://biblehub.com/context/deuteronomy/1-6.htm
deuteronomycontextyahwehgodspoke
https://arxiv.org/html/2510.00615v1
llm agentsaconoptimizingcontextcompression
https://huggingface.co/papers/2408.15542
Join the discussion on this paper page
language modelpaperkangaroopowerfulvideo
https://arxiv.org/abs/2405.15318
Abstract page for arXiv paper 2405.15318: Are Long-LLMs A Necessity For Long-Context Tasks?
longllmsnecessitycontext
https://www.longwarjournal.org/archives/2013/06/missing_context_on_the_tribal.php
Recent coverage of the grassroots uprising against the Taliban in Kandahar province fails to account for a long-running tribal feud.
missingcontexttribalawakeningpanjwai
https://perk-long-context.web.app/
PERK: Long-Context Reasoning as Test-Time Learning
long contextperkreasoningtesttime
https://simonwillison.net/2025/Apr/7/long-context-llm/
LLM 0.24 is now available with new features to help take advantage of the increasingly long input context supported by modern LLMs. (LLM is my command-line...
long contextsupportllmusingfragments
https://www.nationsreportcard.gov/ltt_2012/context_math.aspx
NAEP - 2012 Long-Term Trend: Classroom Context: Coursetaking in Mathematics
long term trendnaepclassroomcontextmathematics
https://openreview.net/forum?id=loJM1acwzf&referrer=%5Bthe%20profile%20of%20Xinyuan%20Lu%5D(%2Fprofile%3Fid%3D~Xinyuan_Lu1)
Understanding documents with rich layouts and multi-modal components is a long-standing and practical task. Recent Large Vision-Language Models (LVLMs) have...
long contextdocument understandingbenchmarkingvisualizationsopenreview
https://openreview.net/forum?id=z5vZDI2r6J&referrer=%5Bthe%20profile%20of%20Hongyin%20Tang%5D(%2Fprofile%3Fid%3D~Hongyin_Tang1)
Processing structured tabular data, particularly large and lengthy tables, constitutes a fundamental yet challenging task for large language models (LLMs)....
large language modelslong contextexploringcapabilitytowards
https://huggingface.co/papers/2503.19325
Join the discussion on this paper page
long contextvideo modelingpaperautoregressivenext
https://openreview.net/forum?id=r0AXK5Cnhr&referrer=%5Bthe%20profile%20of%20Guohao%20Dai%5D(%2Fprofile%3Fid%3D~Guohao_Dai4)
State-of-the-art large language models (LLMs) are now claiming remarkable supported context lengths of 256k or even more. In contrast, the average context...
long contextlvevalbalancedbenchmark
https://openreview.net/forum?id=0xvSZPcZdT&referrer=%5Bthe%20profile%20of%20Volodymyr%20Mnih%5D(%2Fprofile%3Fid%3D~Volodymyr_Mnih1)
In this paper, we present a benchmark to pressure-test today's frontier models' multimodal decision-making capabilities in the very long-context regime (up to...
imitation learningbenchmarkcontextlong
https://openreview.net/forum?id=mMXdHyBcHh&referrer=%5Bthe%20profile%20of%20Jiajie%20Zhang%5D(%2Fprofile%3Fid%3D~Jiajie_Zhang2)
Though current long-context large language models (LLMs) have demonstrated impressive capacities in answering user questions based on extensive text, the lack...
fine grainedenablingllmsgeneratecitations
https://openreview.net/forum?id=FSjIrOm1vz&referrer=%5Bthe%20profile%20of%20Hansi%20Zeng%5D(%2Fprofile%3Fid%3D~Hansi_Zeng1)
The scaling of inference computation has unlocked the potential of long-context large language models (LLMs) across diverse settings. For knowledge-intensive...
long contextinferencescalingretrievalaugmented
https://arxiv.org/abs/2503.01996
Abstract page for arXiv paper 2503.01996: One ruler to measure them all: Benchmarking multilingual long-context language models
onerulermeasurebenchmarkingmultilingual
https://arxiv.org/abs/2502.01549
Abstract page for arXiv paper 2502.01549: VideoRAG: Retrieval-Augmented Generation with Extreme Long-Context Videos
long contextretrievalaugmentedgenerationextreme
https://openreview.net/forum?id=DUsqifwwf5&referrer=%5Bthe%20profile%20of%20Mingbao%20Lin%5D(%2Fprofile%3Fid%3D~Mingbao_Lin1)
Recent advances in long-context large language models (LLMs) make them commercially viable, but their standard attention mechanisms' quadratic complexity...
sparse optimizationsoloslongsequencecontext
https://openreview.net/forum?id=WQwy1rW60F&referrer=%5Bthe%20profile%20of%20Guohao%20Dai%5D(%2Fprofile%3Fid%3D~Guohao_Dai4)
State-of-the-art large language models (LLMs) are now claiming remarkable supported context lengths of 256k or even more. In contrast, the average context...
long contextlvevalbalancedbenchmark
https://aws.amazon.com/blogs/machine-learning/use-the-applyguardrail-api-with-long-context-inputs-and-streaming-outputs-in-amazon-bedrock/
As generative artificial intelligence (AI) applications become more prevalent, maintaining responsible AI principles becomes essential. Without proper...
long contextuseapiinputsstreaming
https://www.biblehub.com/context/john/10-24.htm
johncontextjewsthereforecame