Spring Launch ‘24: Meet DataRobot’s Newest Features to Confidently Build and Deploy Production-Grade GenAI Applications


The most inspiring part of my role is traveling around the globe, meeting our customers from every sector and seeing, learning, collaborating with them as they build GenAI solutions and put them into production. It’s thrilling to see our customers actively advancing their GenAI journey. But many in the market are not, and the gap is growing. 

AI leaders are rightfully struggling to move beyond the prototype and experimental stage, it’s our mission to change that. At DataRobot, we call this the “confidence gap”. It’s the trust, safety and accuracy and concerns surrounding GenAI that are holding teams back, and we are committed to addressing it. And, it’s the core focus of our Spring ’24 launch and its groundbreaking features.

This release focuses on the three most significant hurdles to unlocking value with GenAI. 

First, we’re bringing you enterprise-grade open-source LLM support, and a suite of evaluation and testing metrics, to help you and your teams confidently create production-grade AI applications. To help you safeguard your reputation and prevent risk from AI apps running amok, we’re bringing you real-time intervention and moderation for all your GenAI applications. And finally, to ensure your entire fleet of AI assets stay in peak performance, we’re bringing you a first-of-its-kind multi-cloud and hybrid AI Observability to help you fully govern and optimize all of your AI investments.

Confidently Create Production-Grade AI Applications 

There is a lot of talk about fine-tuning an LLM. But, we have seen that the real value lies in fine-tuning your generative AI application. It’s tricky, though. Unlike predictive AI, which has thousands of easily accessible models and common data science metrics to benchmark and assess performance against, generative AI hasn’t—until now. 

Unlike predictive AI, which has thousands of easily accessible models and common data science metrics to benchmark and assess performance against, generative AI hasn’t—until now.

In our Spring ’24 launch, get enterprise-grade support for any open-source LLM. We’ve also introduced an entire set of LLM evaluation, testing, and metrics. Now, you can fine-tune your generative AI application experience, ensuring their reliability and effectiveness.

Enterprise-Grade Open Source LLMs Hosting

Privacy, control, and flexibility remain critical for all organizations regarding LLMs.There has been no easy answer for AI Leaders who have been stuck with having to pick between vendor lock-in risks using major API-based LLMs that could become sub-optimal and expensive in the immediate future, figuring out how to stand up and host your own open source LLM, or custom-building, hosting, and maintaining your own LLM. 

With our Spring Launch, you have access to the broadest selection of LLMs, allowing you to choose the one that aligns with your security requirements and use cases. Not only do you have ready-to-use access to LLMs from leading providers like Amazon, Google, and Microsoft, but you also have the flexibility to host your own custom LLMs. Additionally, our Spring ’24 Launch offers enterprise-level access to open-source LLMs, further expanding your options.

We have made hosting and using open-source foundational models like LLaMa, Falcon, Mistral, and Hugging Face easy with DataRobot’s built-in LLM security and resources. We have eliminated the complex and labor-intensive manual DevOps integrations required and made it as easy as a drop-down selection.

image
LLM Evaluation, Testing and Assessment Metrics 

With DataRobot, you can freely choose and experiment across LLMs. We also give you advanced experimentation options, such as trying various chunking strategies, embedding methods, and vector databases. With our new LLM evaluation, testing, and assessment metrics, you and your teams now have a clear way of validating the quality of your GenAI application and LLM performance across these experiments. 

With our first-of-its-kind synthetic data generation for prompt-and-answer evaluation, you can quickly and effortlessly create thousands of question-and-answer pairs. This lets you easily see how well your RAG experiment performs and stays true to your vector database.  

We are also giving you an entire set of evaluation metrics. You can benchmark, compare performance, and rank your RAG experiments based on faithfulness, correctness, and other metrics to create high-quality and valuable GenAI applications. 

LLM Evaluation Testing and Assessment Metrics alt
LLM Evaluation Testing and Assessment Metrics

And with DataRobot, it’s always about choice. You can do all of this as low code or in our fully hosted notebooks, which also have a rich set of new codespace functionality that eliminates infrastructure and resource management and facilitates easy collaboration. 

Observe and Intervene in Real-Time

The biggest concern I hear from AI leaders about generative AI is reputational risk. There are already plenty of news articles about GenAI applications exposing private data and legal courts holding companies accountable for the promises their GenAI applications made. In our Spring ’24 Launch, we’ve addressed this issue head-on. 

With our rich library of customizable guards, workflows, and notifications, you can build a multi-layered defense to detect and prevent unexpected or unwanted behaviors across your entire fleet of GenAI applications in real time. 

Our library of pre-built guards can be fully customized to prevent prompt injections and toxicity, detect PII, mitigate hallucinations, and more. Our moderation guards and real-time intervention can be applied to all of your generative AI applications – even those built outside of DataRobot, giving you peace of mind that your AI assets will perform as intended.

Real-time LLM Intervention and Moderation
 Real-time LLM Intervention and Moderation

Govern and Optimize Infrastructure Investments

Because of generative AI, the proliferation of new AI tools, projects, and teams working on them has increased exponentially. I often hear about “shadow GenAI” projects and how AI leaders and IT teams struggle to reign it all in. They find it challenging to get a comprehensive view, compounded by complex multi-cloud and hybrid environments. The lack of AI observability opens organizations up to AI misuse and security risks. 

Cross-Environment AI Observability 

We’re here to help you thrive in this new normal where AI exists in multiple environments and locations. With our Spring ’24 Launch, we’re bringing the first-of-its-kind, cross-environment AI observability –  giving you unified security, governance, and visibility across clouds and on-premise environments. 

Your teams get to work in the tools and ways they want; AI leaders get the unified governance, security, and observability they need to protect their organizations. 

Our customized alerts and notification policies integrate with the tools of your choice, from ITSM to Jira and Slack, to help you reduce time-to-detection (TTD) and time-to-resolution (TTR). 

Insights and visuals help your teams see, diagnose, and troubleshoot issues with your AI assets – Trace prompts to the response and content in your vector database with ease, See Generative AI topic drift with multi-language diagnostics, and more.  

NVIDIA and GPU integrations 

And, if you’ve made investments in NVIDIA, we’re the first and only AI platform to have deep integrations across the entire surface area of NVIDIA’s AI Infrastructure – from NIMS, to NeMoGuard models, to their new Triton inference services, all ready for you at the click of a button. No more managing separate installs or integration points, DataRobot makes accessing your GPU investments easy. 

mNGoFCziYIsZ5oRUmsLtD0lySyaN2QxnW20DBQS2uIH mYs7NJVLIvutVDLjU1csDftTBCUwNtwh9DpFGxcKlf186bQ7sC8a 0Vv8ueZoQS1Opd m57m5T nxJgu8XKRQB9P8zCjTlY ZleRcJd5Ro
Optimized AI Inference and NVIDIA Inference Microservices 

Our Spring ’24 launch is packed with exciting features, including GenAI, predictive capabilities, and enhancements in time series forecasting, multimodal modeling, and data wrangling. 

All of these new features are available in cloud, on-premise, and hybrid environments. So, whether you’re an AI leader or part of an AI team, our Spring ’24 launch sets the foundation for your success. 

This is just the beginning of the innovations we’re bringing you. We have so much more in store for you in the months ahead. Stay tuned as we’re hard at work on the next wave of innovations. 

Get Started 

Learn more about DataRobot’s GenAI solutions and accelerate your journey today. 

  • Join our Catalyst program to accelerate your AI adoption and unlock the full potential of GenAI for your organization.
  • See DataRobot’s GenAI solutions in action by scheduling a demo tailored to your specific needs and use cases.
  • Explore our new features, and connect with your dedicated DataRobot Applied AI Expert to get started with them. 

Join the DataRobot Generative AI Catalyst Program

Accelerate your AI adoption and unlock the full potential of GenAI for your organization


Learn more

About the author

Venky Veeraraghavan
Venky Veeraraghavan

Chief Product Officer

Venky Veeraraghavan leads the Product Team at DataRobot, where he drives the definition and delivery of DataRobot’s AI platform. Venky has over twenty-five years of experience as a product leader, with previous roles at Microsoft and early-stage startup, Trilogy. Venky has spent over a decade building hyperscale BigData and AI platforms for some of the largest and most complex organizations in the world. He lives, hikes and runs in Seattle, WA with his family.


Meet Venky Veeraraghavan