Making an inference call to a Model Endpoint with Open Inference Protocol

Cloudera AI Inference service serves predictive ONNX models using the Nvidia Triton Server. The deployed model endpoints are compliant with Open Interface Protocol version 2.