[UI] Summarization Performance Test Quickstart
Summarization Performance Evaluations on GPT-3.5-Turbo with DynamoEval UI
Last updated: October 10th, 2024
This Quickstart showcases an end-to-end walkthrough of how to utilize DynamoAI’s platform solution to run summarization performance test. It also covers some general guidelines and specific examples for setting up test configurations.
If you are a developer and wish to follow the same quickstart with Dynamo AI’s SDK, we refer to the associated SDK Quickstart.
Create Model
Begin by navigating to the DynamoAI home page. This page contains the model registry – a collection of all the models you have uploaded for evaluation or guardrailing. The model registry contains information to help you identify your model, such as the model source, use case, and date updated.
To upload a new model to the registry, click "Upload new model" button (at the upper right corner). When this button is clicked, a popup will appear, requesting more information about the model.
This includes information such as the model name and use case. The popup will also request information about the model source. Remote inference can be used to create a connection with any model that is provided by a third party or is already hosted and can be accessed through an API endpoint. On the other hand, Local inference can be used to upload a custom model file.
Example. For this quickstart, we recommend setting the following:
- Model name: “GPT-3.5 Summarization”
- Model Source: Remote Inference
The next page of the popup will ask for more detailed information to set up the API connection.
This includes information about the model provider, API access key, model identifier, as well as an optional model endpoint (if required by your API provider).
Example. We recommend setting the following:
- API Provider: OpenAI
- API Key: (your OpenAI API key)
- Model: gpt-3.5-turbo
- Endpoint: (leave blank)
At this point, your model named "GPT3.5 Summarization" should have been created and should be displayed on the models registry.
Create Test
To create a test for the created model, click on "DynamoEval" under Links column from the Model Registry page > Click on "Testing" tab > Click "New test" at the upper right corner.
Time to fill in the test information.
- Fill in the test title to be indicative of the test you are running.
- Select "Performance tests"
- Select the two Test types: “ROUGE” and “BERTScore”.