
Filter papers
Remove All Filters
LLM See, LLM Do: Guiding Data Generation to Target Non-Differentiable Objectives
Generative Models
Language
Tooling
Interpretability
Transformers
Supervised Learning
Open Source
Generative Models
Language
Tooling
Interpretability
Transformers
Supervised Learning
Open Source
Investigating Continual Pretraining in Large Language Models: Insights and Implications
Continual Learning
Transformers
Language
Continual Learning
Transformers
Language
When Less is More: Investigating Data Pruning for Pretraining LLMs at Scale
Transformers
Generative Models
Data Pruning
Large-Scale Pretraining
Data Efficiency
Transformers
Generative Models
Data Pruning
Large-Scale Pretraining
Data Efficiency
Intriguing Properties of Quantization at Scale
Generative Models
Representation Learning
Transformers
Model Compression
Generative Models
Representation Learning
Transformers
Model Compression
Associative Memory Augmented Asynchronous Spatiotemporal Representation Learning for Event-based Perception
Transformers
Representation Learning
Efficiency
Computer Vision
Transformers
Representation Learning
Efficiency
Computer Vision
Exploring Low Rank Training of Deep Neural Networks
Model Compression
Transformers
Efficiency
Model Compression
Transformers
Efficiency
Prioritized Training on Points that are Learnable, Worth Learning, and Not Yet Learnt
Transformers
Representation Learning
Supervised Learning
Unsupervised Learning
Transformers
Representation Learning
Supervised Learning
Unsupervised Learning