I fine-tuned this model on "tool dataset"
#1
by
nikitastaf1996
- opened
As experiment I decided to fine-tune this model on medium-size-generated-tasks dataset.
The goal was to follow ReAct Langchain agent format. While using python_repl tool.
In my experience you need 13b or 30b model to do that.
It successfully follows the format and even tries to write some shitty code.
Given model size and time to fine-tune it's success.
Link:https://huggingface.co/nikitastaf1996/TinyStories-Instruct-33M-react-medium-tasks-dirty
nikitastaf1996
changed discussion title from
I fine-tuned this on "tool dataset"
to I fine-tuned this model on "tool dataset"
nikitastaf1996
changed discussion status to
closed