Channel Avatar

Argilla @UCAIz8TmvQQrLqbD7sd-5S2A@youtube.com

523 subscribers - no pronouns :c

Argilla is a collaboration platform for AI engineers and dom


51:39
Start a token classification project on the Hugging Face Hub with Argilla, GliNER and NuExtract LLM
49:15
Start a text classification project on the Hugging Face Hub with Argilla and SetFit
44:03
Argilla Community Everything image: from fine-tuning CLIP models to synthetic image datasets
00:59
What is distilabel? A brief feature overview.
31:19
Generating and cleaning a preference dataset for DPO / ORPO with LLMs and distilabel
45:54
Optimizing RAG Pipelines by fine-tuning custom embedding models on synthetic data with ZenML
00:32
ZenML a way to streamline your complex projects with ease
00:32
cosine similarity as proxy for quality of sentence pair data
00:39
optimizing RAG by choosing the right model
00:36
model pooling for diverse synthetic data generation
30:12
Ellamind on synthetic data generation with distilabel for pipelining and LLM finetuning
04:45
Scaling Synthetic Data Creation with 1 Billion Personas | PersonaHub Dataset Explained
50:02
Javier Alonso on lead optimisation at Idealista
06:29
Exploring the PRISM Dataset: Conversations, Insights, and Model Performance
39:06
Ben Burtenshaw on the Argilla 2.0 SDK refactor
01:02:59
Louis Guitton on NER with Argilla
54:39
Weights & Biases on Wandbot
32:18
Datamaran on using Argilla in MLOps workflows for ESG governance
39:00
Understanding and reproducing DEITA with MantisNLP using distilabel=1.0.0
38:11
Elad Levi on AutoPrompt and intent-based prompt calibration and prompt engineering
55:51
Daniel van Strien on the Hugging Face hub and synthetic creation of a DPO dataset for Haiku
51:10
Seth Levine on the usage of SetFit and BerTopic for unsupervised clustering
45:40
Red Cross 510 on NLP for good with SetFit for chat message classification
49:29
Prolific on workload distribution, LLM preference data annotation and Phi2 fine-tune Colab
01:03:58
Pitching AI to your boss, SLMs vs LLMs and contributing to open source projects
43:59
Kickstart NLP with synthetic data and running LLMs on Google Colab using vLLM
36:03
How we cleaned OpenBMB UltraFeedback and Notus
36:40
An introduction to distilabel for AI feedback and synthetic data generation
46:16
Deploy Argilla on a private Hugging Face space and how to contribute to open source
45:29
Deploy Argilla on a public Hugging Face space and create multi-modal datasets
02:17
Meet Argilla
00:20
Collect human feedback for evaluating fine-tuned LLMs
00:20
Collect human feedback for fine-tuning ChatGPT models
07:27
Tutorial: Deploy Argilla on Hugging Face and create your first FeebackDataset
21:16
Tutorial: Curate an instruction dataset for supervised fine-tuning
11:45
Guía práctica: ¿Cómo mejorar los datos de Alpaca para entrenar un LLM en Español?
03:00
Argilla: 3 min walkthrough