Remove 2023 Remove AWS Remove Clustering
article thumbnail

PEFT fine tuning of Llama 3 on SageMaker HyperPod with AWS Trainium

AWS Machine Learning Blog

The process of setting up and configuring a distributed training environment can be complex, requiring expertise in server management, cluster configuration, networking and distributed computing. To simplify infrastructure setup and accelerate distributed training, AWS introduced Amazon SageMaker HyperPod in late 2023.

AWS 109
article thumbnail

Racing into the future: How AWS DeepRacer fueled my AI and ML journey

AWS Machine Learning Blog

In 2018, I sat in the audience at AWS re:Invent as Andy Jassy announced AWS DeepRacer —a fully autonomous 1/18th scale race car driven by reinforcement learning. But AWS DeepRacer instantly captured my interest with its promise that even inexperienced developers could get involved in AI and ML.

AWS 109
professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

How climate tech startups are building foundation models with Amazon SageMaker HyperPod

Flipboard

Amazon Web Services (AWS) provides the essential compute infrastructure to support these endeavors, offering scalable and powerful resources through Amazon SageMaker HyperPod. Starting in early 2023, we saw the first wave of climate tech startups adopting generative AI to optimize operations.

AWS 130
article thumbnail

Map Earth’s vegetation in under 20 minutes with Amazon SageMaker

AWS Machine Learning Blog

We pick the first week of December 2023 in this example. By utilizing the search_raster_data_collection function from SageMaker geospatial, we identified 8,581 unique Sentinel-2 images taken in the first week of December 2023. These batches are then evenly distributed across the machines in a cluster. format("/".join(tile_prefix),

ML 123
article thumbnail

Open source observability for AWS Inferentia nodes within Amazon EKS clusters

AWS Machine Learning Blog

Despite the availability of advanced distributed training libraries, it’s common for training and inference jobs to need hundreds of accelerators (GPUs or purpose-built ML chips such as AWS Trainium and AWS Inferentia ), and therefore tens or hundreds of instances. or later NPM version 10.0.0

AWS 129
article thumbnail

AWS at NVIDIA GTC 2024: Accelerate innovation with generative AI on AWS

AWS Machine Learning Blog

AWS was delighted to present to and connect with over 18,000 in-person and 267,000 virtual attendees at NVIDIA GTC, a global artificial intelligence (AI) conference that took place March 2024 in San Jose, California, returning to a hybrid, in-person experience for the first time since 2019.

AWS 136
article thumbnail

Real value, real time: Production AI with Amazon SageMaker and Tecton

AWS Machine Learning Blog

The US nationwide fraud losses topped $10 billion in 2023, a 14% increase from 2022. Orchestrate with Tecton-managed EMR clusters – After features are deployed, Tecton automatically creates the scheduling, provisioning, and orchestration needed for pipelines that can run on Amazon EMR compute engines.

ML 102