From d5d5771971e0851570afb2b3b8e24515b08b4d70 Mon Sep 17 00:00:00 2001 From: idrc-cms-bot <41127776+idrc-cms-bot@users.noreply.github.com> Date: Mon, 16 Sep 2024 16:44:56 -0300 Subject: [PATCH] chore(cms): create "src/collections/resources/researchers-warn-of-unchecked-toxicity-in-ai-language-models.md" (#1992) Automatically generated. Merged on Decap CMS. --- ...nchecked-toxicity-in-ai-language-models.md | 19 +++++++++++++++++++ 1 file changed, 19 insertions(+) create mode 100644 src/collections/resources/researchers-warn-of-unchecked-toxicity-in-ai-language-models.md diff --git a/src/collections/resources/researchers-warn-of-unchecked-toxicity-in-ai-language-models.md b/src/collections/resources/researchers-warn-of-unchecked-toxicity-in-ai-language-models.md new file mode 100644 index 00000000..6356cec5 --- /dev/null +++ b/src/collections/resources/researchers-warn-of-unchecked-toxicity-in-ai-language-models.md @@ -0,0 +1,19 @@ +--- +title: Researchers Warn of Unchecked Toxicity in AI Language Models +focus: AI Ethics/Policy +source: CTV News +readability: + - Intermediate +type: Website Article +openSource: true +link: https://www.ctvnews.ca/sci-tech/researchers-warn-of-unchecked-toxicity-in-ai-language-models-1.6856095 +learnTags: + - machineLearning + - researchCentre + - methods + - bias + - ethics +summary: Researchers from MIT's Improbable AI Lab and the MIT-IBM Watson AI Lab + are developing a “red-team language model” that is designed to generate + problematic prompts that trigger undesirable responses from tested chatbots. +---