Skip to content

Conversation

@sarthakwer
Copy link

Added example "Distillation + Finetuning" under subheading Finetuning. The example distills the capabilities of a large model into a small model using Together's finetuning API. Details: I demonstrate how to use Curator to distill capabilities from a large language model to a much smaller 8B parameter model. I use Yelp restaurant reviews dataset to train a sentiment analysis model. I then generate a synthetic dataset using Bespokelabs's curator and finetune a model using Together's finetuning API. The finetuned model shows a 12% improvement in accuracy while being 14x cheaper than LLM

Added example "Distillation + Finetuning" under subheading Finetuning. The example distills the capabilities of a large model into a small model using Together's finetuning API.  
Details: I demonstrate how to use Curator to distill capabilities from a large language model to a much smaller 8B parameter model. I use Yelp restaurant reviews dataset to train a sentiment analysis model. I then generate a synthetic dataset using Bespokelabs's curator and finetune a model using Together's finetuning API. The finetuned model shows a 12% improvement in accuracy while being 14x cheaper than LLM
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant