Remove en-gb
article thumbnail

Streamline diarization using AI as an assistive technology: ZOO Digital’s story

AWS Machine Learning Blog

SageMaker asynchronous endpoints support upload sizes up to 1 GB and incorporate auto scaling features that efficiently mitigate traffic spikes and save costs during off-peak times. At the time of writing, using the faster-whisper Large V2 model, the resulting tarball representing the SageMaker model is 3 GB in size.

AWS 95
article thumbnail

Accelerate PyTorch with DeepSpeed to train large language models with Intel Habana Gaudi-based DL1 EC2 instances

AWS Machine Learning Blog

billion-parameter model using the wikicorpus-en dataset. Each dl1.24xlarge instance has eight Habana Gaudi accelerators, each with 32 GB of memory and a full mesh RoCE network between cards with a total bi-directional interconnect bandwidth of 700 Gbps each (see Amazon EC2 DL1 instances Deep Dive for more information).

AWS 67
article thumbnail

Fine-tune Llama 2 for text generation on Amazon SageMaker JumpStart

AWS Machine Learning Blog

Fine-tuning technique Language models such as Llama are more than 10 GB or even 100 GB in size. nnn### Explanation:nWe answer the question with the input's date of birth and the date of death.nnn### Solution: 1102n Response from the fine-tuned model: Félix Luna died on November 5th, 2009.nn

ML 97