You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: nemoguardrails/eval/data/moderation/README.md
+2-2Lines changed: 2 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ The jailbreak and output moderation rails can be evaluated using the following c
7
7
The various arguments that can be passed to evaluate the moderation rails are:
8
8
9
9
10
-
- model_name`: Name of the model to use. Default is 'text-davinci-003'.
10
+
-`model_name`: Name of the model to use. Default is 'text-davinci-003'.
11
11
-`llm`: Name of the LLM provide. Default is 'openai'.
12
12
-`dataset-path`: Path to the dataset to evaluate the rails on. The dataset should contain one prompt per line.
13
13
-`split`: The split of the dataset to evaluate on. Choices are 'helpful' or 'harmful'. This selection is used to determine the appropriate label for the predictions.
@@ -54,7 +54,7 @@ By default, we evaluate 50 samples from the dataset. This can be changed using t
54
54
55
55
## Downloading the Anthropic Helpful Dataset
56
56
57
-
The Anthropic Helpful Dataset can be downloaded from the following [link](https://huggingface.co/datasets/Anthropic/hh-rlhf/tree/main/helpful-base).
57
+
The Anthropic Helpful Dataset can be downloaded from the following [link](https://huggingface.co/datasets/Anthropic/hh-rlhf/tree/main/helpful-base).
58
58
59
59
Once it is downloaded, move it to the `nemoguardrails/eval/data/moderation` folder and rename it as ```anthropic_helpful.jsonl.gz```.
0 commit comments