Deepchecks secured $14 million in funding for transforming AI validation

TL;DR:

  • Deepchecks, a leader in AI system validation, unveils its groundbreaking LLM Evaluation solution.
  • This innovative solution addresses the unique challenges posed by Large Language Models (LLMs).
  • It focuses on both the quality of LLM responses and model safety, addressing bias, toxicity, and privacy concerns.
  • Deepchecks’ LLM Evaluation solution offers flexibility to adapt to scenarios with multiple valid LLM responses.
  • Recognizing the diverse user base, it caters to stakeholders such as data curators, product managers, and data scientists.
  • The solution provides tailored evaluation strategies for distinct phases of LLM-based app development.
  • Deepchecks’ CEO, Philip Tannor, believes this solution will expedite the development of LLM-based applications.
  • The company recently secured $14 million in funding through a seed round led by Alpha Wave Ventures, with participation from Hetz Ventures and Grove Ventures.

Main AI News:

In the fast-evolving landscape of artificial intelligence, Deepchecks continues to lead the way with its pioneering solutions. Today, we’re excited to introduce Deepchecks’ revolutionary LLM Evaluation solution, a game-changing advancement poised to redefine the validation of AI systems, particularly those powered by Large Language Models (LLMs).

Since its inception, Deepchecks has been a driving force in the realm of AI system validation. Building upon its highly acclaimed open-source package launched in January 2022 for testing ML models, the company has earned widespread acclaim, boasting a remarkable 3,000 GitHub stars and a staggering 900,000 downloads. Encouraged by the enthusiastic reception from the AI and machine learning community, Deepchecks has expanded its offerings to cater to the diverse requirements of its rapidly growing user base.

The LLM Evaluation solution represents Deepchecks’ response to the escalating demand for effective evaluation tools tailored to LLM-based applications. Deepchecks recognized the inherent challenges posed by LLMs, including the imperative to assess both accuracy and model safety. This entails addressing concerns related to bias, toxicity, and the inadvertent leakage of personally identifiable information (PII). Moreover, the solution acknowledges the unique nature of LLMs, which can yield multiple valid responses for a single input, necessitating flexible testing approaches, including the use of curated “golden sets.”

Key Highlights of Deepchecks’ LLM Evaluation Solution:

  1. Dual Focus: Our solution places equal emphasis on evaluating the quality of LLM responses in terms of accuracy, relevance, and usefulness, while also ensuring robust model safety by addressing bias, toxicity, and strict adherence to privacy policies.
  2. Flexible Testing: Recognizing the variability in LLM responses, our solution adapts to scenarios where multiple valid outputs may be generated from a single input, offering the essential flexibility required for comprehensive testing.
  3. Diverse User Base: Deepchecks acknowledges that LLM-based applications involve collaboration and input from various stakeholders, including data curators, product managers, business analysts, data scientists, and machine learning engineers. Our solution caters to the needs of this diverse user base.
  4. Phased Approach: We understand the distinct phases in LLM-based app development, from the experimental and developmental stages to staging, beta testing, and, ultimately, production. Deepchecks provides tailored evaluation strategies for each phase, ensuring a seamless journey toward production-ready applications.

Philip Tannor, CEO at Deepchecks, emphasizes the significance of the LLM Evaluation solution, stating, “Companies in the market are swiftly creating ‘quick-and-dirty’ proof-of-concepts based on APIs like OpenAI and prompt engineering. However, the transition to production-ready applications is often hindered by challenges related to quality, consistency, and policy compliance. We firmly believe that our LLM Evaluation solution can significantly expedite the development of LLM-based applications while ensuring their safety.”

In a testament to Deepchecks’ commitment to innovation, the company recently secured a substantial $14 million in funding through a seed round. This investment, led by Alpha Wave Ventures, with contributions from Hetz Ventures and Grove Ventures, underscores the industry’s recognition of Deepchecks’ potential to revolutionize AI system validation.

Conclusion:

Deepchecks’ LLM Evaluation solution represents a significant leap in AI system validation, addressing critical challenges and providing flexibility for LLM-based applications. With substantial funding and strong industry recognition, Deepchecks is poised to make a significant impact on the AI market, offering a safer and more efficient path to production-ready LLM-powered solutions.

Source