人気の記事一覧
In-Context Retrieval-Augmented Language Models
You Only Cache Once: Decoder-Decoder Architectures for Language Models
MemLLM: Finetuning LLMs to Use An Explicit Read-Write Memory
State-Free Inference of State-Space Models: The Transfer Function Approach
Ferret-v2: An Improved Baseline for Referring and Grounding with Large Language Models
MoEUT: Mixture-of-Experts Universal Transformers
Lessons from the Trenches on Reproducible Evaluation of Language Models
Scaling Transformer to 1M tokens and beyond with RMT
Thinking Tokens for Language Modeling
Granite Code Models: A Family of Open Foundation Models for Code Intelligence
Text summarization with ChatGPT for drug labeling documents
Infini-gram: Scaling Unbounded n-gram Language Models to a Trillion Tokens
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
On the Long Range Abilities of Transformers
Towards Graph Foundation Models: A Survey and Beyond
Transformers are Multi-State RNNs
Fewer Truncations Improve Language Modeling
X-LoRA: Mixture of Low-Rank Adapter Experts, a Flexible Framework for Large Language Models with Applications in Protein Mechanics and Design