Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
devass
devass121
Follow
0 followers
·
2 following
AI & ML interests
None yet
Recent Activity
updated
a model
about 8 hours ago
devass121/kkroto
replied
to
ezgikorkmaz
's
post
5 days ago
If you are interested in reinforcement learning now you can register for the AAAI 2025 Tutorial I am organizing! Link: https://sites.google.com/view/aisafety-aaai2025
replied
to
alielfilali01
's
post
5 days ago
~75% on the challenging GPQA with only 40M parameters 🔥🥳 GREAT ACHIEVEMENT ! Or is it ? This new Work, "Data Laundering: Artificially Boosting Benchmark Results through Knowledge Distillation", take out the mystery about many models i personally suspected their results. Speacially on leaderboards other than the english one, Like the Open Arabic LLM Leaderbaord https://huggingface.co/spaces/OALL/Open-Arabic-LLM-Leaderboard. The authors of this work, first started by training a model on the GPQA data, which, unsurprisingly, led to the model achieving 100% performance. Afterward, they trained what they referred to as a 'legitimate' model on legitimate data (MedMCQA). However, they introduced a distillation loss from the earlier, 'cheated' model. What they discovered was fascinating: the knowledge of GPQA leaked through this distillation loss, even though the legitimate model was never explicitly trained on GPQA during this stage. This raises important questions about the careful use of distillation in model training, especially when the training data is opaque. As they demonstrated, it’s apparently possible to (intentionally or unintentionally) leak test data through this method. Find out more: https://huggingface.co/papers/2412.15255
View all activity
Organizations
None yet
models
3
Sort: Recently updated
devass121/kkroto
Text-to-Image
•
Updated
about 8 hours ago
•
devass121/Airing
Updated
11 days ago
devass121/JewelAi
Updated
11 days ago
datasets
None public yet