H2O Driverless AI offers model deployment, management and monitoring capabilities for the IT and DevOps teams. The platform makes it convenient for IT to deploy the winning model across a broad range of production environments. In addition, the data science and data engineering teams can monitor the performance of the model for any drifts in predictions and scores over time, as well as manage any re-training or tuning necessary at run-time.

Operationalizing Models

As enterprises “make their own AI”, new challenges emerge:

  • Maintaining reproducibility, traceability, and verifiability of machine learning models
  • Recording experiments, tracking insights, reproducibility of results
  • Searchability of models (or querying models)
  • Collaborating within the DS and ML teams
  • Visualizing model performance (drift, degradation, A/B testing)

Operationalizing models crosses functional boundaries

  • DevOps and IT teams are usually heavily involved
  • Model operations should require minimal changes to existing application workflows
  • Maintain data and model lineage in case of rollbacks, regulatory compliance

Model Deployment Options with Driverless AI

Driverless AI offers the following options for deploying machine learning (ML) models, depending on where the AI application is running:

  • 1. As a cloud service
  • 2. As a local REST endpoint
  • 3. As a Java object file

• The model could be directly deployed in a cloud service.

• Driverless AI offers the ability to export the model directly in AWS Lambda or Sagemaker.

• Ideal for building and running your AI applications in AWS.

• Configuration details can be seen here.

• The model could be configured to run on a local REST server.

• Driverless AI offers the ability to deploy the scoring pipeline on a local server.

• Ideal for AI workloads in on-premises environments.

• Configuration details can be seen here.

• The model could be abstracted into a Java object as a standalone model scoring engine.

• Driverless AI allows downloading the model as a Plain Old Java Object (POJO) or Model Object Optimized (MOJO) file.

• POJO and MOJO files are standalone scoring engines. Ideal for running AI applications in low-latency environments such as edge devices or on-premises.

• Configuration details can be seen here.

Model Operations

As AI applications enter production:

  • Data scientists and data engineers need to manage the transition of ML models
  • Development → Staging → Production
  • Model accuracy can drift over time. Data scientists need alerts if drift exceeds certain thresholds.
  • DevOps/IT need a central store for models, model artifacts, related inference, etc.
  • ML engineers may need to re-calibrate or re-tune production models
  • Seamless collaboration between data science, DevOps and IT teams becomes important

Key MLOps Capabilities in Driverless AI Today

Model Management

(Data Science Teams)

  • Projects Workspace
  • Model store
  • Collaboration

Model Deployment


  • Deploy to different environments – in the cloud, on-premises

Model Monitoring

(Business/Data Science)

  • Specify metrics or parameters to monitor
MLOps with H2O Driverless AI

Model Monitoring and Management Done Right

Driverless AI includes new capabilities for model administration, monitoring and management. These capabilities allow:

• DevOps teams to monitor the models for system health checks

• Data science teams to monitor metrics around drift detection, model degradation, A/B testing

• Provides alerts for recalibration and retraining

Telemetry and Alerts

Driverless AI can monitor models for drift, anomalies, model metrics and residuals, and provide alerts on a dashboard for potential re-tuning or re-training of models.

Model comparison

The MLOps offers capabilities to compare multiple models by looking at their confusion matrices. Moreover, our A/B testing functionality helps to run and compare multiple models in production before they are deployed in production.

Full model traceability

Data scientists can track back a prediction on a specific model and investigate the report to understand how it was created. This also includes the ability to frequently retrain and publish updated models to the runtime environment.

Model Governance

It is crucial for IT and DevOps to implement access rules and decision rights across the enterprises as AI goes into production. MLOps provides important capabilities such as role-based access controls for models as well as tracking who built the model and who deployed it.

Deploy Model as a REST Endpoint

Start Your 21-Day Free Trial Today

Get It Now
Desktop img