Relari

Relari

Software Development

San Francisco, California 452 followers

Data-driven toolkit to evaluate and improve LLM applications.

About us

Relari is the only data-driven toolkit to evaluate and improve LLM applications. AI developers leverage Relari to take the guesswork out of LLM development and ship reliable AI products faster. We are backed by top investors from Y Combinator, Soma Capital and General Catalyst. We are hiring! Please reach out!

Website
https://www.relari.ai
Industry
Software Development
Company size
2-10 employees
Headquarters
San Francisco, California
Type
Privately Held
Founded
2023
Specialties
LLM, AI, GenerativeAI, LLMOps, Infrastructure, EnterpriseAI, EvaluationPipeline, Experimentation, and Reliability

Locations

Employees at Relari

Updates

  • Relari reposted this

    View profile for Yi Zhang, graphic

    Co-Founder @ Relari (YC W24) | Scale your LLM evaluation and ship faster

    Thank you Qdrant for being a great partner! David and Thierry did a great job walking through how to use Relari for end-to-end RAG evaluation and optimization. This case study also includes a sneak peek to our latest feature: Auto Prompt Optimization (APO). Many companies using APO have seen significant improvement in prompt performance. Ping me if you'd like to give it a try! #RAG #PromptEngineering #LLM #EnterpriseAI

    View organization page for Qdrant, graphic

    25,871 followers

    How do you measure the performance of a RAG app? We’ve been working with Relari to make that easier. In our latest blog, Thierry Damiba, David Myriel, and Yi Zhang show practical steps for: ✅ Adjusting 𝐓𝐨𝐩-𝐊 𝐩𝐚𝐫𝐚𝐦𝐞𝐭𝐞𝐫𝐬 to return the right number of results ✅ Using 𝐀𝐮𝐭𝐨 𝐏𝐫𝐨𝐦𝐩𝐭 𝐎𝐩𝐭𝐢𝐦𝐢𝐳𝐚𝐭𝐢𝐨𝐧 to refine chatbot responses ✅ Evaluating it all using the GitLab legal policies dataset as an example With this workflow, you can run fast, iterative tests, experiment with hybrid search, and access detailed metrics like precision, recall, and rank-based evaluations. Check out the full blog for code and implementation 👉 https://lnkd.in/dsw3BQrx

    • No alternative text description for this image
  • Relari reposted this

    View organization page for Qdrant, graphic

    25,871 followers

    How do you measure the performance of a RAG app? We’ve been working with Relari to make that easier. In our latest blog, Thierry Damiba, David Myriel, and Yi Zhang show practical steps for: ✅ Adjusting 𝐓𝐨𝐩-𝐊 𝐩𝐚𝐫𝐚𝐦𝐞𝐭𝐞𝐫𝐬 to return the right number of results ✅ Using 𝐀𝐮𝐭𝐨 𝐏𝐫𝐨𝐦𝐩𝐭 𝐎𝐩𝐭𝐢𝐦𝐢𝐳𝐚𝐭𝐢𝐨𝐧 to refine chatbot responses ✅ Evaluating it all using the GitLab legal policies dataset as an example With this workflow, you can run fast, iterative tests, experiment with hybrid search, and access detailed metrics like precision, recall, and rank-based evaluations. Check out the full blog for code and implementation 👉 https://lnkd.in/dsw3BQrx

    • No alternative text description for this image
  • Relari reposted this

    View profile for Yi Zhang, graphic

    Co-Founder @ Relari (YC W24) | Scale your LLM evaluation and ship faster

    Building an LLM application? It's time to rethink how you approach evaluation. - "We’re too early for evaluation.” - "We don’t have enough resources to do this now." - "😅, I just eyeball it.” Too often I hear these comments in my conversation with LLM developers. But these are missed opportunities. Evaluation isn't just about avoiding mistakes—it’s a powerful tool to differentiate your product from the noise. We see it over and over again with our customers: AI teams/startups that embrace evaluation from the start don’t just catch errors—they learn, iterate, and build better products, faster. It’s not a luxury; it’s a necessity. “You can’t improve what you can’t measure.” A strong evaluation framework is more than a safety net; it’s your secret weapon to winning in a crowded market. So, what’s holding you back from running evaluation? Curious about how strong evaluation can elevate your product? Let’s discuss! #LLMEvaluation #EnterpriseAI #RAG #LLMAgents

  • View organization page for Relari, graphic

    452 followers

    Discover how Vanta's AI team launched multiple successful LLM products, leveraging Relari's data-driven evaluation and improvement toolkit!

    View profile for Yi Zhang, graphic

    Co-Founder @ Relari (YC W24) | Scale your LLM evaluation and ship faster

    Few things make me happier than seeing our customers flourish and bring powerful AI applications to life.   Last week, Vanta raised $150 million at a $2.45 billion valuation, further solidifying its lead in the security compliance space.   Over the last several months, we've been working closely with Vanta's AI team. We’ve witnessed firsthand how effectively they leveraged a data-driven approach to productionize several high-impact LLM products. The result? Overwhelming positive feedback from Vanta's customers!   To date, Relari has proudly supported Vanta with over 50,000 evaluation runs on various synthetic datasets, ensuring robust performance for each LLM product.   A key recurring trend we've observed: contrary to popular belief, extensive testing accelerates iteration speed rather than slowing it down. Teams that implement data-driven tests from the prototyping phase reach production faster. They achieve this by making quick, confident decisions backed by rapid, targeted experiments on key components such as RAG parameters, prompts and LLM architectural choices.   Congratulations to Vanta on this milestone! We're excited to see the innovative Generative AI products they'll launch next.   Check out more details in this case study on our approach and results!   What trends have you noticed among teams building successful LLM products? What strategies seem to resonate most with customers? Share your insights in the comments! #LLM #RAG #GenerativeAI #LLMEvaluation #DataDrivenDevelopment

    How Vanta Leverages Relari.ai for Enhanced LLM-Powered Compliance Product

    How Vanta Leverages Relari.ai for Enhanced LLM-Powered Compliance Product

    relari.ai

  • Relari reposted this

    View profile for Yi Zhang, graphic

    Co-Founder @ Relari (YC W24) | Scale your LLM evaluation and ship faster

    Excited to collaborate with Qdrant on this webinar next week. We will walk through some of the best practices to build production-grade RAG systems.

    View organization page for Qdrant, graphic

    25,871 followers

    Another exciting opportunity to learn advanced RAG testing methods! Yi Zhang from Relari will be joining Thierry Damiba to discuss how synthetic data can be used to create production-ready RAG applications. This is a great opportunity for data scientists, ML engineers and developers working on GenAI to calibrate and improve their projects. The webinar will be followed up by a Q&A session, so bring your best questions and topics to discuss. See you on July 31st!

    This content isn’t available here

    Access this content and more in the LinkedIn app

  • Relari reposted this

    View organization page for Qdrant, graphic

    25,871 followers

    Another exciting opportunity to learn advanced RAG testing methods! Yi Zhang from Relari will be joining Thierry Damiba to discuss how synthetic data can be used to create production-ready RAG applications. This is a great opportunity for data scientists, ML engineers and developers working on GenAI to calibrate and improve their projects. The webinar will be followed up by a Q&A session, so bring your best questions and topics to discuss. See you on July 31st!

    This content isn’t available here

    Access this content and more in the LinkedIn app

  • Relari reposted this

    View profile for Pasquale Antonante, graphic

    PhD @ MIT | Co-Founder @ Relari AI (YC W24)

    "Is RAG all you need?" - check out this talk by Sara Zanzottera at EuroPython, where she highlighted our integration with Haystack, the popular LLM framework by deepset.   Some highlights: 1️⃣ How RAG works and how it fails 2️⃣ ChatGPT (without RAG) vs Perplexity (with RAG) 3️⃣ How to use continuous-eval by Relari to evaluate RAG performance objectively   Check out the presentation, slides, and example code in this link: https://lnkd.in/erqypcAr

    • No alternative text description for this image
  • Relari reposted this

    Are you getting the most out of your LLMs and RAGs? In the latest episode of the ODSC Ai X Podcast, join Pasquale Antonante, PhD, Co-founder and CTO at Relari AI, to discuss innovative evaluation methods for LLM and RAG applications. Get ready to explore: ✅ Measuring LLM Performance: The different ways to evaluate LLM pipelines  ✅ Understanding RAGs: Key evaluation metrics and how to measure precision, recall, faithfulness, relevance, and correctness ✅ The Human in the Loop: The role human evaluation plays in assessing the quality of generated text ✅ The Role of Synthetic Data: Using synthetic data based evaluation for LLMs and RAG and leveraging synthetic data for LLM fine-tuning 🎧 Listen Now:  https://linktr.ee/odsc #AI #LargeLanguageModels #RAG  #Podcast #ODSC

    • No alternative text description for this image
  • Relari reposted this

    View profile for Yi Zhang, graphic

    Co-Founder @ Relari (YC W24) | Scale your LLM evaluation and ship faster

    I was inspired by the SF community breakfasts hosted by Inferless and started organizing "GenAI Developer Breakfast" at Harvard Square while we are in Boston. I find these intimate small-group conversations super fun and insightful. Thank you, everyone, for coming today and exchanging learnings on building LLM-powered applications across a variety of industries! Jon Duffy, Babak Kia, Chris Tierney, Axel Solano, umamaheswar E., Tyce Herrman, Gabor Soter

    • No alternative text description for this image
  • Relari reposted this

    View profile for Pasquale Antonante, graphic

    PhD @ MIT | Co-Founder @ Relari AI (YC W24)

    Earlier this month, I had a great conversation with Sheamus from Open Data Science Conference (ODSC). We discussed RAG evaluation, the challenges of evaluating complex pipelines and how to effectively leverage synthetic-data.  I also shared some of our latest learnings from working with top AI teams deploying LLMs in production. Our chat is out now in the latest Ai X podcast episode! Check it out here: 🍏 Apple Podcast (https://hubs.li/Q02Cvbdx0) 🎧 SoundCloud (https://hubs.li/Q02CvbXF0) 🎵 Spotify (https://hubs.li/Q02Cv9Wb0) 📦 Castbox (https://hubs.li/Q02CvcmW0) Reach out to me if you have any questions or want to learn more about our work at Relari. #AI #MachineLearning #DataScience #ODSC #AiX #RAG #Evaluation #LLM #GenAI #SyntheticData

    ‎ODSC's Ai X Podcast: How to Evaluate Large Language Models and RAG Applications with Pasquale Antonante on Apple Podcasts

    ‎ODSC's Ai X Podcast: How to Evaluate Large Language Models and RAG Applications with Pasquale Antonante on Apple Podcasts

    podcasts.apple.com

Similar pages

Funding

Relari 1 total round

Last Round

Pre seed

US$ 500.0K

Investors

Y Combinator
See more info on crunchbase