-
TokDrift: When LLM Speaks in Subwords but Code Speaks in Grammar
Paper • 2510.14972 • Published • 34 -
LightMem: Lightweight and Efficient Memory-Augmented Generation
Paper • 2510.18866 • Published • 110 -
Every Attention Matters: An Efficient Hybrid Architecture for Long-Context Reasoning
Paper • 2510.19338 • Published • 114 -
The Smol Training Playbook
📚2.7kThe secrets to building world-class LLMs
Jonatan Borkowski PRO
j14i
AI & ML interests
None yet
Recent Activity
liked
a Space
about 4 hours ago
hysts/daily-papers
reacted
to
sergiopaniego's
post
with 🚀
about 4 hours ago
Google DeepMind releases FunctionGemma, a 240M model specialized in 🔧 tool calling, built for fine-tuning
TRL has day-0 support. To celebrate, we’re sharing 2 new resources:
> Colab guide to fine-tune it for 🌐 browser control with BrowserGym OpenEnv
> Standalone training script
> Colab notebook: https://colab.research.google.com/github/huggingface/trl/blob/main/examples/notebooks/grpo_functiongemma_browsergym_openenv.ipynb
> Training script: https://github.com/huggingface/trl/blob/main/examples/scripts/openenv/browsergym_llm.py (command to run it inside the script)
> More notebooks in TRL: https://huggingface.co/docs/trl/example_overview#notebooks