Skip to content

deepchecks/llm-evals-demos

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 

Repository files navigation

Deepchecks LLM Evaluation - Demos

Hands-on demos showing how to evaluate, debug, and monitor LLM apps with Deepchecks.

Demos

Demo Description Framework
Content Creator Crew Evaluate and debug a multi-agent content creation workflow CrewAI

About Deepchecks

Deepchecks is a comprehensive solution for AI evaluations, helping you make sure your AI applications work properly from PoC to production.

Deepchecks LLM Evaluation is for testing, validating and monitoring LLM-based applications. With Deepchecks you can continuously validate LLM-based applications including characteristics, performance metrics, and potential pitfalls throughout the entire lifecycle from pre-deployment and internal experimentation to production.

Deepchecks' platform streamlines configuring auto-scoring and applying it to your LLM-based app during dev, CI/CD and production, making LLM evaluations achivable at scale.

Explore the full documentation.

Links

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors