Deprecated : The each() function is deprecated. This message will be suppressed on further calls in /home/zhenxiangba/zhenxiangba.com/public_html/phproxy-improved-master/index.php on line 456
context - a leonardlin Collection
leonardlin 's Collections context updated Aug 14, 2024
Soaring from 4K to 400K: Extending LLM's Context with Activation Beacon Paper
• 2401.03462
• Published Jan 7, 2024 • 29
MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers Paper
• 2305.07185
• Published May 12, 2023 • 10
YaRN: Efficient Context Window Extension of Large Language Models Paper
• 2309.00071
• Published Aug 31, 2023 • 83
Infinite-LLM: Efficient LLM Service for Long Context with DistAttention
and Distributed KVCache Paper
• 2401.02669
• Published Jan 5, 2024 • 17
LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning Paper
• 2401.01325
• Published Jan 2, 2024 • 27
Extending Context Window of Large Language Models via Semantic
Compression Paper
• 2312.09571
• Published Dec 15, 2023 • 16
Zebra: Extending Context Window with Layerwise Grouped Local-Global
Attention Paper
• 2312.08618
• Published Dec 14, 2023 • 13
Mamba: Linear-Time Sequence Modeling with Selective State Spaces Paper
• 2312.00752
• Published Dec 1, 2023 • 150
E^2-LLM: Efficient and Extreme Length Extension of Large Language Models Paper
• 2401.06951
• Published Jan 13, 2024 • 26
Extending LLMs' Context Window with 100 Samples Paper
• 2401.07004
• Published Jan 13, 2024 • 16
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache
Quantization Paper
• 2401.18079
• Published Jan 31, 2024 • 8
KIVI: A Tuning-Free Asymmetric 2bit Quantization for KV Cache Paper
• 2402.02750
• Published Feb 5, 2024 • 5
A Human-Inspired Reading Agent with Gist Memory of Very Long Contexts Paper
• 2402.09727
• Published Feb 15, 2024 • 38
In Search of Needles in a 10M Haystack: Recurrent Memory Finds What LLMs
Miss Paper
• 2402.10790
• Published Feb 16, 2024 • 42
LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper
• 2402.13753
• Published Feb 21, 2024 • 116
Data Engineering for Scaling Language Models to 128K Context Paper
• 2402.10171
• Published Feb 15, 2024 • 25
Striped Attention: Faster Ring Attention for Causal Transformers Paper
• 2311.09431
• Published Nov 15, 2023 • 4
Ring Attention with Blockwise Transformers for Near-Infinite Context Paper
• 2310.01889
• Published Oct 3, 2023 • 13
LLoCO: Learning Long Contexts Offline Paper
• 2404.07979
• Published Apr 11, 2024 • 22
Leave No Context Behind: Efficient Infinite Context Transformers with
Infini-attention Paper
• 2404.07143
• Published Apr 10, 2024 • 111
InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding
Extremely Long Sequences with Training-Free Memory Paper
• 2402.04617
• Published Feb 7, 2024 • 6
Megalodon: Efficient LLM Pretraining and Inference with Unlimited
Context Length Paper
• 2404.08801
• Published Apr 12, 2024 • 66
TransformerFAM: Feedback attention is working memory Paper
• 2404.09173
• Published Apr 14, 2024 • 43
Extending Llama-3's Context Ten-Fold Overnight Paper
• 2404.19553
• Published Apr 30, 2024 • 34
Make Your LLM Fully Utilize the Context Paper
• 2404.16811
• Published Apr 25, 2024 • 55
Long-context LLMs Struggle with Long In-context Learning Paper
• 2404.02060
• Published Apr 2, 2024 • 37
HyperAttention: Long-context Attention in Near-Linear Time Paper
• 2310.05869
• Published Oct 9, 2023 • 2
World Model on Million-Length Video And Language With RingAttention Paper
• 2402.08268
• Published Feb 13, 2024 • 40
LongNet: Scaling Transformers to 1,000,000,000 Tokens Paper
• 2307.02486
• Published Jul 5, 2023 • 82
MInference 1.0: Accelerating Pre-filling for Long-Context LLMs via
Dynamic Sparse Attention Paper
• 2407.02490
• Published Jul 2, 2024 • 26
LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs Paper
• 2408.07055
• Published Aug 13, 2024 • 69