Channel Avatar

Arize AI @UCrVHzD-psX5IMCGoEWHXmGw@youtube.com

4.2K subscribers - no pronouns :c

Arize AI is an AI observability and evaluation platform, fro


05:47
Phoenix: Guardrails AI Integration Walkthrough
05:48
Phoenix: Function Call Evaluations
07:31
Phoenix: Experiments Walkthrough
06:44
Phoenix: Datasets
04:12
Introducing: Hosted Phoenix
06:39
Introducing: Arize Copilot
04:01
Platform demo 2024
04:48
Guardrails for AI Applications
09:12
Text To SQL: Datasets and Experiments
10:49
LLM Evaluation: Getting Started
03:47
Tracing a Multimodal Query Application
11:01
Running Custom LLM Evaluations ⚙️: Function Calling Agent
05:12
LLM App AB Testing Using Projects
44:22
Community Paper Reading: RAFT – Adapting Language Model to Domain Specific RAG.
09:54
LLM Evals and LLM as a Judge: Fundamentals
03:06
LLM Token Counting
03:38
LLM Tracing: Getting Started
09:10
Debugging Sessions in LLM Chatbot Applications with Tracing and Evals
45:55
LLM Interpretability: Exploring the Latest Research from OpenAI and Anthropic
11:08
Understanding Embeddings Similarity Search In Production Workflows
49:14
Trustworthy LLMs: a Survey and Guideline for Evaluating Large Language Models' Alignment
56:28
Improving & Evaluating Your RAG Application with MongoDB and Arize Phoenix
47:09
LLM Evals for Router Based Architectures
46:10
SQL Generation Evals: LLMs-as-a-Judge
44:31
Breaking Down EvalGen: Who Validates the Validators?
32:30
Optimizing AI Apps for Real-World Performance with OpenAI and Flipkart
43:15
Keys to Understanding ReAct: Synergizing Reasoning and Acting in Language Models
44:14
LLM Evaluation In Practice: Timeseries Evals
46:02
LLM Evals In Practice: Creating Custom Task Evals
47:19
Towards A Production-Ready Customer Feedback LLM: Leveraging Evals for Advanced Feedback Analysis
58:43
Optimizing LLM Retrieval Strategies with Arize AI & Qdrant
44:34
Chronos: Learning the Language of Time Series
53:47
LLM Evaluation Essentials: Benchmarking and Analyzing Retrieval Approaches
46:46
LLM Evaluation Essentials: Statistical Analysis of Summarization LLM Evaluations
02:12
Arize onboarding: Dashboards for LLM Use Case
04:28
Arize onboarding: Generative Embedding
01:10
Arize onboarding: LLM Data Ingestion
05:47
Arize Onboarding: LLM Tracing
04:32
Arize Onboarding: Prompt Playground
02:02
Arize Onboarding: Data Ingestion
04:55
Arize Onboarding: Exploring Dashboards
06:29
Arize Onboarding: Performance Tracing
04:10
Arize Onboarding: Setting Monitors for Unstructured Use Case
07:07
Arize Onboarding: NLP Embeddings
02:26
Arize Onboarding: Uploading Models and Data
42:57
Anthropic Claude 3: A GPT-4 Competitor Has Arrived
44:51
Reinforcement Learning in the Era of LLMs
45:16
Understanding OpenAI's Sora & Evaluating Large Video Model Generation
39:53
RAG vs Fine-tuning
44:22
RAG Time! Evaluate RAG with LLM Evals and Benchmarking
44:35
Phi-2 Model
14:06
Arize Product Demo
05:07
Phoenix OSS: AI Observability & Evaluation in a Notebook
29:23
LangChain RAG featuring Shopify's Madhav Thaker
35:10
Vibe-Based Prompt Engineering with PromptLayer's Jared Zoneraich
35:26
Troubleshooting and Evaluating LLMs with Amber Roberts
47:47
Constructing an Evaluation Approach for Generative AI Models with Hugging Face's Rajiv Shah
47:56
Mistral AI (Mixtral-8x7B): Performance, Benchmarks
01:16:47
Fine Tuning and Evaluating LLMs with Anyscale and Arize
50:54
LLM Evaluation Essentials: Statistical Analysis of Hallucination LLM Evaluations