Burning ray

adarksky
·

AI & ML interests

None yet

Recent Activity

liked a model 2 days ago
deepseek-ai/DeepSeek-R1
updated a model 7 days ago
hexgrad/Kokoro-82M
new activity 7 days ago
hexgrad/Kokoro-82M:Update kokoro.py
View all activity

Organizations

fast.ai community's profile picture Hugging Face Discord Community's profile picture

adarksky's activity

New activity in hexgrad/Kokoro-82M 7 days ago

Update kokoro.py

#43 opened 7 days ago by
adarksky
reacted to merve's post with 🔥 about 2 months ago
view post
Post
2667
small but mighty 🔥
you can fine-tune SmolVLM on an L4 with batch size of 4 and it will only take 16.4 GB VRAM 🫰🏻 also with gradient accumulation simulated batch size is 16 ✨
I made a notebook that includes all the goodies: QLoRA, gradient accumulation, gradient checkpointing with explanations on how they work 💝 https://github.com/huggingface/smollm/blob/main/finetuning/Smol_VLM_FT.ipynb