Remove artifact-evaluation
article thumbnail

No "Zero-Shot" Without Exponential Data

Hacker News

Web-crawled pretraining datasets underlie the impressive "zero-shot" evaluation performance of multimodal models, such as CLIP for classification/retrieval and Stable-Diffusion for image generation.

139
139
article thumbnail

A defined process for project post mortem review (1996)

Hacker News

The authors propose a tentative, standard process for conducting post mortem reviews and describe activities, roles, and artifacts of the process. Participants are empowered when they know that each issue raised during the post mortem process must be added to the risk database and evaluated methodically on each subsequent project.

Database 131
professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Are Model Explanations Useful in Practice? Rethinking How to Support Human-ML Interactions.

ML @ CMU

Our work further motivates novel directions for developing and evaluating tools to support human-ML interactions. For example, explanations are thought to assist model developers in identifying when models rely on spurious artifacts and to aid domain experts in determining whether to follow a model’s prediction.

ML 246
article thumbnail

How VMware built an MLOps pipeline from scratch using GitLab, Amazon MWAA, and Amazon SageMaker

Flipboard

It is critical for the VMware Carbon Black team to design and build a custom end-to-end MLOps pipeline that orchestrates and automates workflows in the ML lifecycle and enables model training, evaluations, and deployments. The following architecture diagram illustrates the end-to-end workflow and the components involved in our MLOps pipeline.

ML 102
article thumbnail

Are Model Explanations Useful in Practice? Rethinking How to Support Human-ML Interactions.

ML @ CMU

Our work further motivates novel directions for developing and evaluating tools to support human-ML interactions. For example, explanations are thought to assist model developers in identifying when models rely on spurious artifacts and to aid domain experts in determining whether to follow a model’s prediction.

ML 130
article thumbnail

Demystifying deepfake videos: The powerful fusion of technology and data science

Data Science Dojo

Synthetic Artifacts: Look for strange artifacts or distortions in the video, such as unnatural lighting, inconsistent shadows, or pixelation. Audio Discrepancies: With the rise of its audio, it is essential to consider auditory cues when evaluating media authenticity. These anomalies can help identify potential fakes.

article thumbnail

Streamline diarization using AI as an assistive technology: ZOO Digital’s story

AWS Machine Learning Blog

In this collaboration, we deployed and evaluated WhisperX on SageMaker, using an asynchronous inference endpoint to host the model. In the following sections, we delve into the details of deploying the WhisperX model on SageMaker, and evaluate the diarization performance. __dict__[WAV2VEC2_MODEL].get_model(dl_kwargs={"model_dir":

AWS 95