Fine-tune Mixtral 8x7b on AWS SageMaker and Deploy to RunPod
Mlearning.ai
DECEMBER 22, 2023
Deploy Fine-tuned Mixtral on RunPod The last step in this tutorial is to deploy the fine-tuned version of Mixtral 8x7b and we chose RunPod (not a sponsor) because of the easy availability and overall good pricing for on-demand GPUs. Upload model to huggingface hub To upload our fine-tuned model to huggingface, we must now create a repository.
Let's personalize your content