What's New
Major features and updates for the Cloudera AI data service.
December 17, 2025
Release notes and fixed issues for version 2.0.53-b273.
New Features / Improvements
- Added a new and dedicated Cloudera AI Overview page in the Control Plane. This page now serves as the primary landing page, replacing the previous Workspace Administration panel view. (DSE-43180)
- Model endpoints now support the configuration of vetted environment variables using the UI and API. This enables specialized use cases, such as toggling reasoning output for NVIDIA NIMs. (DSE-45592)
- Added support for the AWS Bahrain region. (DSE-46039)
- Added support for the Azure Spain region. (DSE-44839)
- Added monitoring for metadata reporting, including the tracking of the workspace version (appVersion). (DSE-32676)
- Added Non-Transparent Proxy (NTP) support. (DSE-39437)
Fixed Issues
- Upgraded NIM-CLI to the latest version to incorporate improvements in memory management and error handling when downloading large models. (DSE-46374)
- All critical and high-severity CVEs in Cloudera AI Registry have been addressed.
- The Registered Models UI now loads successfully when using Google Chrome version 142 or above. (DSE-48824)
- The Model Endpoints UI now loads successfully when using Google Chrome version 142 or above. (DSE-48824)
- Resolved a critical race condition in the reconciler that caused successfully completed jobs to be incorrectly marked as failed with a -1 exit code. (DSE-48076 and DSE-44901)
-
The Spark UI button is now correctly displayed in the dropdown menu for Jupyter Notebooks within the Cloudera AI/ML environment. (DSE-42562)
- Resolved an issue where users with the Team Admin role were unable to update team descriptions due to a permission mismatch. The system now correctly aligns the required permission levels, allowing team creators to manage their team settings as expected. (DSE-48770)
Behavioral Change
- Increased file transfer timeouts to 10 minutes for uploads and 15 minutes for downloads, ensuring reliable API v2 transfer of large files. (DSE-39671)
Deprecation Notice
- NVIDIA optimized profiles for the following models are no longer supported and have been
removed from the Model Hub in this release:
- Llama 3.2 11B
- Llama 3.2 90B
- Mixtral 8x22B
Although these optimized profiles are no longer available in the Model Hub, the models remain supported and accessible through Hugging Face.
October 31, 2025
Release notes and fixed issues for version 2.0.53-b241.
New Features / Improvements
- Improved model information display by surfacing attributes like MaxTokens, Parameters, and Dimensions for Embedding Models, enabling better decision-making before importing models.
- Added the ability for users to start and stop deployed model endpoints, providing greater control over resource management and cost optimization. This feature allows you to pause inactive endpoints to save resources while maintaining the ability to quickly restart them when needed.
- Improved user experience by enabling customers to open model endpoints in different browser tabs, allowing for better multitasking and simultaneous monitoring of multiple endpoints.
- Enhanced API accessibility by providing customers access to the Swagger UI for invoking AI Inference APIs, enabling interactive API testing and documentation exploration.
- Added visibility to the underlying compute cluster through a direct link in the UI for AI Inference instances, providing seamless navigation to cluster details and infrastructure monitoring.
- Implemented validation for root volume size when adding new nodes to AI Inference deployments, preventing configuration errors and ensuring adequate storage capacity.
- Enhanced node group management by displaying GPU types and enabling filtering based on GPU type when searching through available nodes, improving resource selection efficiency.
- Upgraded KServe from version 0.12 to 0.15, enhancing the underlying model serving infrastructure.
- Enhanced support for self-signed certificates in public cloud deployments, resolving installation and update pain points.
- Re-enabled cadence workflow for workbench upgrades, fixing version compatibility issues.
- Optimized the loading speed of the Site Administration overview and the Projects pages by improving cluster-wide resource usage data collection, ensuring quick loading even in environments with 1,000+ user namespaces.
- The Job Retry feature introduces automated recovery for failed, timed out, or skipped jobs. Administrators can now set concurrent retry limits and customize the retry behavior to enhance job resilience and eliminate the need for manual failure intervention.
Fixed Issues
- Fixed Events & Logs rendering during the creation of new registries in the UI. (DSE-48101)
- Resolved inconsistencies in the tooltips displayed on the Model Endpoint Details page, ensuring accurate and helpful information is shown to users. (DSE-45042)
- The API service can now inspect the metadata of MLflow models from the Cloudera AI Registry to correctly identify the Hugging Face transformer flavor. This ensures the model is automatically routed to the correct runtime (i.e. HuggingFace runtime with vLLM backend unlike defaulting to Triton + ONNX earlier), enabling successful deployment of MLflow transformer models. (DSE-46462)
- Removed the rigid requirement for every team to have an Administrator, allowing Team Administrators (or Site Administrators if none are designated) to manage team membership and roles. (DSE-41700)
- You can now customize the model build process with two new options: model_root_dir allows setting a custom source directory, and build_script_path enables specifying a custom location for the build script. (DSE-45166)
- Resolved a page layout issue that occurred after horizontal window resizing. The fix ensures the browser view restores correctly without requiring a manual refresh. (DSE-40393).
- Fixed an issue where system messages were hidden or suppressed when a Workload (such as a Session, Job, or Application) failed to start up while using a PBJ Workbench Runtime. (DSE-46126)
Behavioral Change
- The R kernel output mechanism has been updated to improve visibility of interactive commands. The output of R commands (for example, install.packages("sparklyr")) is now mirrored to both the Session tab and the Logs tab by default, rather than appearing solely in the Logs. This ensures immediate visibility of important messages like installation progress, compilation output, and errors. Users can revert to the previous behavior by setting the environment variable JUPYTER_KERNEL_OUTPUT_FILTER_REGEX to "DISABLED" or "^$". (DSE-35335)
- Clearing a session now permanently removes all content from the session history, ensuring that subsequent PDF exports and email reports contain only current, relevant results and no stale data from previous runs. (DSE-12565)
- MLFlow integration with transformer models is now complete, enabling models saved using the workbench (via MLFlow) in the Hugging Face transformer format to be successfully deployed on Cloudera AI Inference service, including both regular and finetuned models. (DSE-43840)
September 25, 2025
Release notes and fixed issues for version 2.0.52-b<TBA>.
New Features / Improvements
- You can now update the visibility of models directly from the UI.
- You can now delete registries in a failed status directly from the UI.
- Added the ability to use pre-downloaded artifacts when importing external models.
- The model card for embedding models now includes
MaxTokens, Parameters,andDimensions. - Added support for upgrading Azure registries with UDR-enabled subnets.
-
Improved error resolution suggestions have been added to the UI for Cloudera AI Registry, empowering users to troubleshoot issues on their own.
- Added support for new models, including NeMo Retriever’s GraphicElements, PageElements, and TableStructure, as well as PaddleOCR, Boltz-2, and GPT-OSS models.
- The Test Model feature is now available for Speech-to-Text models.
- You can now specify vLLM arguments when deploying Hugging Face models.
- Improved error resolution suggestions have been added to the UI for Cloudera AI Inference service empowering users to troubleshoot issues on their own.
- Grouped CPU/GPU nodes under the Cloudera AI Inference service Details page for a more user-friendly experience.
- You can now provision multiple CPU/GPU resource groups in the Cloudera AI Workbench. This provides enhanced control over workload scheduling and allows segregation of workloads based on the instance types.
-
Added support for EKS 1.32.
-
Added support for AKS 1.32.
Fixed Issues
- Addressed critical CVEs in most NGC models by upgrading to the latest NIM versions. (DSE-47435)
-
Resolved an issue where code samples were not correctly rendered for riva, reranking, and retrieval models. (DSE-46413)
-
Ensured that the UI carries out proper validation of GPU, CPU, and memory during model deployment. (DSE-46250)
-
Resolved an issue where Prometheus errors were not surfaced in the Model Endpoints UI. (DSE-46077)
-
Addressed an accessibility issue with the 'Deploy external dropdown button'. (DSE-46071)
-
Resolved an issue that blocked navigation to other tabs while waiting for a response from the Model Endpoint's Test Model. (DSE-46247)
-
Resolved an issue where a scale alert box was not displayed in the UI when the model endpoint was ready and the current replica count was zero. (DSE-46301)
- Resolved an issue where the UI did not properly surface failures occurring during NVIDIA model imports from the Model Hub page. (DSE-45971)
- Resolved an issue where the UI was not correctly auto-selecting the running registry. (DSE-46246)
- Resolved an issue that prevented users from registering models from within a workbench. (DSE-47077)
- Resolved an issue where proxy environment variables were being overwritten during model builds. (DSE-46070)
- Resolved an issue where the Project Settings page displayed the project owner's ID instead of their name. (DSE-46572)
- Resolved an issue where spaces were incorrectly added to the ldap_dn during synchronization of some users, which caused those members of CML groups to be unintentionally removed. (DSE-47591)
