What's New

Major features and updates for the Cloudera Machine Learning data service.

June 20, 2024

Release notes and fixed issues for version 2.0.45-b82.

New Features / Improvements

  • You can now delete the Model Endpoint under AI Inference from the Cloudera Machine Learning UI.

Fixed Issues

  • Previously, the Model Registry service was broken due to a change in the certificates. This issue is now resolved.

June 11, 2024

Release notes and fixed issues for version 2.0.45-b81.

Fixed Issues

  • Previously, users were unable to upgrade their model registry or create new model registries if their cluster was resized. This issue is now resolved.

May 29, 2024

Release notes and fixed issues for version 2.0.45-b81.

New Features / Improvements

  • The NodeSelector label can now be added for inference services. The label can be specified in the instance_type field of the deploy or update requests. This would enable you to direct inference service pods to specific nodes.
  • Enhancements to the Export API to support the Observability APIs.

Fixed Issues

  • Added support to create model registries in AWS with a public load balancer from the UI. Previously, this option was available only from the backend.

May 15, 2024

Release notes and fixed issues for version 2.0.45-b76.

New Features / Improvements

  • Cloudera AI Inference Service (Technical Preview): AI Inference service is a production-grade serving environment for traditional, generative AI, and LLM models. It is designed to handle the challenges of production deployments, such as high availability, fault tolerance, and scalability. The service is now available for users to carry out inference on the following three categories of models:
    • TRT-LLMs: LLMs that are optimized to TRT engine and available in NVIDIA GPU Cloud catalog, also known as NGC catalog.
    • LLMs available through Hugging Face Hub.
    • Traditional machine learning models like classification, regression, and so on.

    Models need to be imported to the model registry to be served using the Cloudera AI Inference Service.

    For more information, see Cloudera AI Inference Service.

  • Cloudera Copilot (Technical Preview): Cloudera Copilot is a highly configurable AI coding assistant integrated with the JupyterLab editor. The Copilot improves developer productivity by debugging code, answering questions and generate notebooks.

    For more information, see Cloudera Copilot.

  • Model Registry API (Technical Preview): New API is available from the Model Registry service to import, get, update and delete models without relying on the CML Workspace service.

    For more information, see Model Registry API.

  • Ephemeral storage limit: The default ephemeral storage limit for CML Projects has been increased from 10 GB to 30 GB.

Fixed Issues

  • Fixed an error that occurs while sorting public projects on the Jobs column.
  • Fixed a bug that was uploading files to the root directory of a project instead to the specified subfolder.