volodia
- Karma
- 545
- Created
- 17 years ago
Recent Submissions
- 1. ▲ Mercury 2 on PinchBench: Diffusion LLM benchmarked on real OpenClaw agent tasks (inceptionlabs.ai)
- 2. ▲ Mercury 2: Best-in-class speed-optimized intelligence at 1,200 tok/SEC (twitter.com)
- 3. ▲ Finetuning 3-Bit LLMs on Consumer GPUs by Integrating with Modular Quantizers (arxiv.org)
- 4. ▲ LLMTune: 4-Bit finetuning of 65B LLAMA models on a single consumer GPU (github.com)
- 5. ▲ LLMTune: 4-Bit Finetuning of LLMs on a Consumer GPI (twitter.com)
- 6. ▲ Don't have a $5k MacBook to run LLAMA65B? MiniLLM runs LLMs on GPUs in <500 LOC (github.com)