Source code for pyrit.datasets.librAI_do_not_answer_dataset

# Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.

from datasets import load_dataset

from pyrit.models import SeedPromptDataset
from pyrit.models.seed_prompt import SeedPrompt


[docs] def fetch_librAI_do_not_answer_dataset() -> SeedPromptDataset: """ Fetch the LibrAI 'Do Not Answer' dataset and return it as a SeedPromptDataset. Relevant Columns: - "risk_area" - "types_of_harm" - "specific_harms" - "question" Returns: SeedPromptDataset: A SeedPromptDataset containing the examples. Note: For more info - Paper - https://arxiv.org/abs/2308.13387 - Github - https://github.com/libr-ai/do-not-answer - HF Dataset - https://huggingface.co/datasets/LibrAI/do-not-answer """ # Load dataset from Hugging Face data = load_dataset("LibrAI/do-not-answer", split="train") seed_prompts = [ SeedPrompt( value=entry["question"], data_type="text", name="", dataset_name="LibrAI/Do-Not-Answer", harm_categories=[entry["risk_area"], entry["types_of_harm"], entry["specific_harms"]], description=( f"This is a prompt from the 'Do Not Answer' dataset under the risk area: {entry['risk_area']}, " f"harm type: {entry['types_of_harm']}, and specific harm: {entry['specific_harms']}." ), source="https://huggingface.co/datasets/LibrAI/do-not-answer", ) for entry in data ] # Create a SeedPromptDataset from the list of SeedPrompt instances return SeedPromptDataset(prompts=seed_prompts)