Deprecated : The each() function is deprecated. This message will be suppressed on further calls in /home/zhenxiangba/zhenxiangba.com/public_html/phproxy-improved-master/index.php on line 456
Papers - a segmond Collection
Papers updated Feb 18, 2024
Chain-of-Thought Reasoning Without Prompting Paper
• 2402.10200
• Published Feb 15, 2024 • 109
How to Train Data-Efficient LLMs Paper
• 2402.09668
• Published Feb 15, 2024 • 43
BitDelta: Your Fine-Tune May Only Be Worth One Bit Paper
• 2402.10193
• Published Feb 15, 2024 • 21
A Human-Inspired Reading Agent with Gist Memory of Very Long Contexts Paper
• 2402.09727
• Published Feb 15, 2024 • 38
Self-Play Fine-Tuning Converts Weak Language Models to Strong Language
Models Paper
• 2401.01335
• Published Jan 2, 2024 • 69
OS-Copilot: Towards Generalist Computer Agents with Self-Improvement Paper
• 2402.07456
• Published Feb 12, 2024 • 46
BiLLM: Pushing the Limit of Post-Training Quantization for LLMs Paper
• 2402.04291
• Published Feb 6, 2024 • 50
Self-Discover: Large Language Models Self-Compose Reasoning Structures Paper
• 2402.03620
• Published Feb 6, 2024 • 117
Shortened LLaMA: A Simple Depth Pruning for Large Language Models Paper
• 2402.02834
• Published Feb 5, 2024 • 17
TrustLLM: Trustworthiness in Large Language Models Paper
• 2401.05561
• Published Jan 10, 2024 • 69
SliceGPT: Compress Large Language Models by Deleting Rows and Columns Paper
• 2401.15024
• Published Jan 26, 2024 • 73
DeepSeek-Coder: When the Large Language Model Meets Programming -- The
Rise of Code Intelligence Paper
• 2401.14196
• Published Jan 25, 2024 • 72
Spotting LLMs With Binoculars: Zero-Shot Detection of Machine-Generated
Text Paper
• 2401.12070
• Published Jan 22, 2024 • 45
Self-RAG: Learning to Retrieve, Generate, and Critique through
Self-Reflection Paper
• 2310.11511
• Published Oct 17, 2023 • 80
Chain-of-Verification Reduces Hallucination in Large Language Models Paper
• 2309.11495
• Published Sep 20, 2023 • 40
Adapting Large Language Models via Reading Comprehension Paper
• 2309.09530
• Published Sep 18, 2023 • 82
Language Modeling Is Compression Paper
• 2309.10668
• Published Sep 19, 2023 • 85
Paper
• 2309.16609
• Published Sep 28, 2023 • 38
CodeFusion: A Pre-trained Diffusion Model for Code Generation Paper
• 2310.17680
• Published Oct 26, 2023 • 74
Extending LLMs' Context Window with 100 Samples Paper
• 2401.07004
• Published Jan 13, 2024 • 16
The Impact of Reasoning Step Length on Large Language Models Paper
• 2401.04925
• Published Jan 10, 2024 • 18
Paper
• 2401.04088
• Published Jan 8, 2024 • 160
Alpha-CLIP: A CLIP Model Focusing on Wherever You Want Paper
• 2312.03818
• Published Dec 6, 2023 • 34
Chain of Code: Reasoning with a Language Model-Augmented Code Emulator Paper
• 2312.04474
• Published Dec 7, 2023 • 34
ZipLoRA: Any Subject in Any Style by Effectively Merging LoRAs Paper
• 2311.13600
• Published Nov 22, 2023 • 47
The Generative AI Paradox: "What It Can Create, It May Not Understand" Paper
• 2311.00059
• Published Oct 31, 2023 • 19
CodePlan: Repository-level Coding using LLMs and Planning Paper
• 2309.12499
• Published Sep 21, 2023 • 80
LoRAShear: Efficient Large Language Model Structured Pruning and
Knowledge Recovery Paper
• 2310.18356
• Published Oct 24, 2023 • 24
Agents: An Open-source Framework for Autonomous Language Agents Paper
• 2309.07870
• Published Sep 14, 2023 • 43
Direct Language Model Alignment from Online AI Feedback Paper
• 2402.04792
• Published Feb 7, 2024 • 35
Rethinking Interpretability in the Era of Large Language Models Paper
• 2402.01761
• Published Jan 30, 2024 • 23
PDFTriage: Question Answering over Long, Structured Documents Paper
• 2309.08872
• Published Sep 16, 2023 • 55
OLMo: Accelerating the Science of Language Models Paper
• 2402.00838
• Published Feb 1, 2024 • 85
Self-Rewarding Language Models Paper
• 2401.10020
• Published Jan 18, 2024 • 153
ReFT: Reasoning with Reinforced Fine-Tuning Paper
• 2401.08967
• Published Jan 17, 2024 • 32
Understanding LLMs: A Comprehensive Overview from Training to Inference Paper
• 2401.02038
• Published Jan 4, 2024 • 65
LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning Paper
• 2401.01325
• Published Jan 2, 2024 • 27
A Comprehensive Study of Knowledge Editing for Large Language Models Paper
• 2401.01286
• Published Jan 2, 2024 • 21
Time is Encoded in the Weights of Finetuned Language Models Paper
• 2312.13401
• Published Dec 20, 2023 • 20
TinyGSM: achieving >80% on GSM8k with small language models Paper
• 2312.09241
• Published Dec 14, 2023 • 40
Llama Guard: LLM-based Input-Output Safeguard for Human-AI Conversations Paper
• 2312.06674
• Published Dec 7, 2023 • 8
Magicoder: Source Code Is All You Need Paper
• 2312.02120
• Published Dec 4, 2023 • 82
Using Human Feedback to Fine-tune Diffusion Models without Any Reward
Model Paper
• 2311.13231
• Published Nov 22, 2023 • 28
Exponentially Faster Language Modelling Paper
• 2311.10770
• Published Nov 15, 2023 • 119
Orca 2: Teaching Small Language Models How to Reason Paper
• 2311.11045
• Published Nov 18, 2023 • 77
Fast Chain-of-Thought: A Glance of Future from Parallel Decoding Leads
to Answers Faster Paper
• 2311.08263
• Published Nov 14, 2023 • 16
OpenMathInstruct-1: A 1.8 Million Math Instruction Tuning Dataset Paper
• 2402.10176
• Published Feb 15, 2024 • 38
Generative Representational Instruction Tuning Paper
• 2402.09906
• Published Feb 15, 2024 • 54
Prometheus: Inducing Fine-grained Evaluation Capability in Language
Models Paper
• 2310.08491
• Published Oct 12, 2023 • 57
Tuna: Instruction Tuning using Feedback from Large Language Models Paper
• 2310.13385
• Published Oct 20, 2023 • 10
AgentTuning: Enabling Generalized Agent Abilities for LLMs Paper
• 2310.12823
• Published Oct 19, 2023 • 36
LoRA Fine-tuning Efficiently Undoes Safety Training in Llama 2-Chat 70B Paper
• 2310.20624
• Published Oct 31, 2023 • 13
Learning From Mistakes Makes LLM Better Reasoner Paper
• 2310.20689
• Published Oct 31, 2023 • 29
S-LoRA: Serving Thousands of Concurrent LoRA Adapters Paper
• 2311.03285
• Published Nov 6, 2023 • 30
Levels of AGI for Operationalizing Progress on the Path to AGI Paper
• 2311.02462
• Published Nov 4, 2023 • 36
Can LLMs Follow Simple Rules? Paper
• 2311.04235
• Published Nov 6, 2023 • 13
LLaMA Pro: Progressive LLaMA with Block Expansion Paper
• 2401.02415
• Published Jan 4, 2024 • 54
A Zero-Shot Language Agent for Computer Control with Structured
Reflection Paper
• 2310.08740
• Published Oct 12, 2023 • 15
Premise Order Matters in Reasoning with Large Language Models Paper
• 2402.08939
• Published Feb 14, 2024 • 28
More Agents Is All You Need Paper
• 2402.05120
• Published Feb 3, 2024 • 57
Infinite-LLM: Efficient LLM Service for Long Context with DistAttention
and Distributed KVCache Paper
• 2401.02669
• Published Jan 5, 2024 • 17
Supervised Knowledge Makes Large Language Models Better In-context
Learners Paper
• 2312.15918
• Published Dec 26, 2023 • 9
Instruction-tuning Aligns LLMs to the Human Brain Paper
• 2312.00575
• Published Dec 1, 2023 • 15
Prompt Engineering a Prompt Engineer Paper
• 2311.05661
• Published Nov 9, 2023 • 23
MFTCoder: Boosting Code LLMs with Multitask Fine-Tuning Paper
• 2311.02303
• Published Nov 4, 2023 • 12
Tell Your Model Where to Attend: Post-hoc Attention Steering for LLMs Paper
• 2311.02262
• Published Nov 3, 2023 • 14
Personas as a Way to Model Truthfulness in Language Models Paper
• 2310.18168
• Published Oct 27, 2023 • 5
Improving Text Embeddings with Large Language Models Paper
• 2401.00368
• Published Dec 31, 2023 • 82
Customizing Language Model Responses with Contrastive In-Context
Learning Paper
• 2401.17390
• Published Jan 30, 2024 Aya Model: An Instruction Finetuned Open-Access Multilingual Language
Model Paper
• 2402.07827
• Published Feb 12, 2024 • 48
Efficient Tool Use with Chain-of-Abstraction Reasoning Paper
• 2401.17464
• Published Jan 30, 2024 • 21
Specialized Language Models with Cheap Inference from Limited Domain
Data Paper
• 2402.01093
• Published Feb 2, 2024 • 47
Can Mamba Learn How to Learn? A Comparative Study on In-Context Learning
Tasks Paper
• 2402.04248
• Published Feb 6, 2024 • 32
Tag-LLM: Repurposing General-Purpose LLMs for Specialized Domains Paper
• 2402.05140
• Published Feb 6, 2024 • 23
TeacherLM: Teaching to Fish Rather Than Giving the Fish, Language
Modeling Likewise Paper
• 2310.19019
• Published Oct 29, 2023 • 9
Language Models can be Logical Solvers Paper
• 2311.06158
• Published Nov 10, 2023 • 20
GPQA: A Graduate-Level Google-Proof Q&A Benchmark Paper
• 2311.12022
• Published Nov 20, 2023 • 36
Memory Augmented Language Models through Mixture of Word Experts Paper
• 2311.10768
• Published Nov 15, 2023 • 19
Digital Socrates: Evaluating LLMs through explanation critiques Paper
• 2311.09613
• Published Nov 16, 2023 • 1
On the Prospects of Incorporating Large Language Models (LLMs) in
Automated Planning and Scheduling (APS) Paper
• 2401.02500
• Published Jan 4, 2024 • 1
In-Context Principle Learning from Mistakes Paper
• 2402.05403
• Published Feb 8, 2024 • 18
Can Large Language Models Understand Context? Paper
• 2402.00858
• Published Feb 1, 2024 • 24
Data Engineering for Scaling Language Models to 128K Context Paper
• 2402.10171
• Published Feb 15, 2024 • 25
A Closer Look at the Limitations of Instruction Tuning Paper
• 2402.05119
• Published Feb 3, 2024 • 5
CodeIt: Self-Improving Language Models with Prioritized Hindsight Replay Paper
• 2402.04858
• Published Feb 7, 2024 • 15
Code Representation Learning At Scale Paper
• 2402.01935
• Published Feb 2, 2024 • 13