---
title: Draft and run prompts
short: Draft and run prompts
tier: enterprise
type: guide
order: 0
order_enterprise: 231
meta_title: Draft your Prompt
meta_description: Create and evaluate an LLM prompt
section: Prompts
date: 2024-06-12 14:09:09
---
With your [Prompt created](prompts_create), you can begin drafting your prompt content to run against baseline tasks.
## Draft a prompt and generate predictions
1. Select your base model.
The models that appear depend on the [API keys](prompts_keys) that you have configured for your organization.
2. In the **Prompt** field, enter your prompt. Keep in mind the following:
* You must include the text variables. These appear directly above the prompt field. (In the demo below, this is the `review` variable.) Click the text variable name to insert it into the prompt.
* Although not strictly required, you should provide definitions for each class to ensure prediction accuracy and to help [add context](#Add-context).
!!! info Tip
You can generate an initial draft by simply adding the text variables and then [clicking **Enhance Prompt**](#Enhance-prompt).
3. Select your baseline:
* **All Project Tasks** - Generate predictions for all tasks in the project. Depending on the size of your project, this might take some time to process. This does not generate an accuracy score for the prompt.
See the [Bootstrapping projects with prompts](prompts_overview#Bootstrapping-projects-with-Prompts) use case.
* **Sample Tasks** - Generate predictions for the first 20 tasks in the project. This does not generate an accuracy score for the prompt.
See the [Bootstrapping projects with prompts](prompts_overview#Bootstrapping-projects-with-Prompts) use case.
* **Ground Truths** - Generate predictions and a prompt accuracy score for all tasks with ground truth annotations. This option is only available if your project has ground truth annotations.
See the [Auto-labeling with Prompts](prompts_overview#Auto-labeling-with-Prompts) use case and the [Prompt evaluation and fine-tuning](prompts_overview#Prompt-evaluation-and-fine-tuning).
4. If this is your first version of the prompt or you want to adjust the current version, click **Save**.
If you want to create a new version of the prompt so that you can compare evaluations between versions, click the drop-down menu next to **Save** and select **Save As**.
5. Click **Evaluate** (if running against a ground truth baseline) or **Run**.
!!! warning
When you click **Evaluate** or **Run**, you will create predictions for each task in the baseline you selected and overwrite any previous predictions you generated with this prompt.
Evaluating your Prompts can result in multiple predictions on your tasks: if you have multiple Prompts for one Project, or if you click both **Evaluate**/**Run** and **Get Predictions for All Tasks from a Prompt**, you will see multiple predictions for tasks in the Data Manager.
## Evaluation results
When you evaluate a prompt, you will see the following metrics:
| Metric | Tasks | Config | Description |
|---|---|---|---|
| **Overall accuracy** | Ground Truths | All labeling configs | A measure of how many predictions are correct when measured against the ground truth. For example, if there are 10 ground truths and your Prompt's predictions match 7 of them, then the overall accuracy would be `0.70`. |
| **Outputs** | All task types | All labeling configs | Number of tasks evaluated. |
| **F1 Score** | Ground Truths |
Single ` |
The [F1 score](https://en.wikipedia.org/wiki/F-score) is a metric to assess a machine learning model's accuracy. This is measured in terms of **precision** and **recall** as follows: `F1 = 2 * (precision * recall) / (precision + recall)` * **Precision**: The proportion of correct positive predictions out of all positive predictions. This is a measure of the quality of the predictions, and looks at how many predictions are correctly aligned with ground truths. This metric is useful for limiting false positives. It answers the question: "How often were the positive predictions correct?" or "How much can I trust a positive prediction?" In other words, it measures how good the model is at not making false predictions. * **Recall**: The proportion of correct positive predictions out of all actual positives. This is a measure of a model's ability to find all positive predictions, and is especially useful in situations in which multiple labels are possible. This metric is useful for maximizing the number positive predictions that you get back. It answers the question "How many of the actual positives were successfully identified?" In other words, it measures how good the model is at not missing things. Note that a "positive" prediction denotes either a 'positive' label (like a checkbox), or the presence of a particular choice/label in the prediction. |
| **Inference cost** | All task types | All labeling configs | The cost to run the prompt based on the number of tokens required. |