Introduction
Cerebrium’s fine-tuning functionality is in public beta and so we are adding more functionality each week! Currently, we only support the training of CasualLM models. If you have an urgent requirement, we can help you just reach out to support
The fine-tuning functionality on Cerebrium allows you to quickly and conveniently fine-tune your LLMs on Cerebrium with just one line of code. Cerebrium leverages the latest techniques such as PEFT and LoRA in order to train models in order to do so in the shortest amount of time (and therefore cost) while still achieving the same performance.
Currently, our fine-tuning capabilities are limited to any causal language models that support 8bit quantisation/LoRA from the HuggingFace transformers library. Some of these models include:
You can use any size variations of the models above. Additionally, we currently use the Alpaca-Lora format as the prompt template format. In future, we plan to release capabilities for custom prompting templates.
We recommend you look at our guide on how to best curate your dataset in order to maximize the performance of your model and make sure it can handle sufficient edge cases.