Evaluating generated datasets for fine-tuning LLMs
Learn how to evaluate synthetic data generated in previous steps using the Large Language Model as the evaluator.
-
In the Cloudera
console, click the Cloudera AI
tile.
The Cloudera AI Workbenches page displays.
-
Click on the name of the workbench.
The workbenches Home page displays.
- Click AI Studios.
- Under AI Studios, click Synthetic Data Studio.
- Under Evaluate, click Getting Started. The Evaluator page is displayed.
- In Display Name, provide a name for the evaluation result.
- In Prompt, enter your custom prompt, or click Restore Default Prompt to reset to the default prompt.
- In Model Provider, select either AWS Bedrock or Cloudera AI Inference as a model provider.
- In Model, select your desired model for the evaluation.
- In Evaluation Examples, update the justification text and score provided by the LLM as needed.
- Under Parameters, modify the model parameters for Temperature, Top K, and Top P to customize the model’s behavior.
-
Click Evalute to start the evaluation.
After completing the evaluation, you can access the generated evaluation report. Click
on the Actions column to re-evaluate the dataset, view the evaluation results, or remove the evaluation report.