Settings

Theme

codelion

Karma
3,049
Created
18 years ago

Recent Submissions

  1. 1. Scaling Pedagogical Pre-Training: From Optimal Mixing to 10B Tokens (huggingface.co)
  2. 2. From HashHop to Memory-Augmented Language Models (huggingface.co)
  3. 3. The Optimal Architecture for Small Language Models (huggingface.co)
  4. 4. Enhancing LLMs with LoRA – Standardized Recipes for Capability Enhancement (huggingface.co)
  5. 5. OpenEvolve: Teaching LLMs to Discover Algorithms Through Evolution (algorithmicsuperintelligence.ai)
  6. 6. Ellora: Enhancing LLMs with LoRA Standardized Recipes for Capability Enhancement (huggingface.co)
  7. 7. The 1B Token Challenge: Finding the Perfect Pre-Training Mix (huggingface.co)
  8. 8. The 1B Token Challenge: Finding the Perfect Pre-Training Mix (huggingface.co)
  9. 9. Pivotal Token Search (PTS): Targeting Critical Decision Points in LLM Training (huggingface.co)
  10. 10. Internal Coherence Maximization(ICM): Label-Free Unsupervised Training Framework (github.com)

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection