GovLoop

How MLOps Helps Agencies Get the Most Out of AI

Building a large AI data model from scratch is prohibitively expensive for most organizations. But the release of AI foundation models, particularly LLMs, has allowed organizations to take advantage of broader community investments in AI.

But there’s a catch: Agencies need to tweak and train these models to meet their specific objectives, maintain privacy and ensure regulatory compliance.

“These models are trained on a broad range of data, a wide understanding of language and concepts and images,” said John Dvorak, Chief Technology Officer, Red Hat North America Public Sector. “Fine-tuning is essential to maximizing the relevance of the model, the accuracy, the effectiveness for the specific use cases that you have.”

Understanding Foundation Models

Keep in mind that foundation models are purposefully versatile and adaptable. Because they draw from a broad base of inputs, “they can be adapted for use across a wide variety of ranges and use cases,” Dvorak said.

But that strength is also a potential weakness. In their raw form, broad-based foundation models are not ideally suited to support the nuanced needs of government agencies, which deal in “unique libraries or vocabularies, terminologies and processes, aren’t necessarily captured or semantically linked in the model. They are also not tuned to address bias or handle concepts such as novel or emerging topics.”

Those models also may not be adept at protecting sensitive or private data, or may not act in accordance with agency regulations and collection authorities.

How MLOps Helps

By using their own data to adjust the foundation model, whether through fine-tuning or newer strategies such as Retrieval-Augmented Generation (RAG), agencies can drive more effective AI outputs. But any effort to maintain a model in production requires a secure, transparent and consistent process for making improvements over time.

This is where Machine Learning Operations (MLOps) comes into the picture. MLOps offers a streamlined approach to making iterative improvements to the model. “It’s taking that model through its life cycle: from data collection, to training that data, putting it into production and then monitoring — then going back and doing it again,” Dvorak said.

Red Hat OpenShift AI provides a flexible and scalable platform for building AI-enabled applications. It includes all the elements of MLOps, empowering organizations to automate and simplify the iterative process of integrating ML models into software development processes, production rollout, monitoring, retraining and redeployment for continued accuracy.

With a flexible platform built on a scalable infrastructure, developers can hone the foundation models in support of AI applications that understand government’s highly specific subject matter and align to its particular operational requirements.

“OpenShift AI can run on prem, in the cloud or on the edge of your network. The platform is plug-and- play: it’s consistent, it’s flexible and provides all the components” to train, fine-tune, serve and monitor models, Dvorak said.

This article appears in our guide, “Getting Practical with AI.” For more examples of how agencies are making real-world use of AI technology, download it here:

 

 

Photo by Noah Wilke at Pexels.com
Exit mobile version