From 8623ca0f3eb5fb4e7bed533b934424cd6d781cd9 Mon Sep 17 00:00:00 2001 From: pandora <128635000+pandora-s-git@users.noreply.github.com> Date: Wed, 4 Dec 2024 15:32:01 +0100 Subject: [PATCH 1/2] URL Typo Fix --- docs/capabilities/guardrailing.mdx | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/capabilities/guardrailing.mdx b/docs/capabilities/guardrailing.mdx index 36e1266..527902b 100644 --- a/docs/capabilities/guardrailing.mdx +++ b/docs/capabilities/guardrailing.mdx @@ -183,8 +183,8 @@ The table below describes the types of content that can be detected in the moder ### Cookbook -Our [moderation cookbook](https://colab.research.google.com/github.com/mistralai/cookbook/tree/main/mistral/moderation/system-level-guardrails.ipynb) provides a concrete example of how to use the Moderation service to implement system level guardrails. -For a more broad view, there is also a more [explorative cookbook](https://colab.research.google.com/github.com/mistralai/cookbook/tree/main/mistral/moderation/moderation-explored.ipynb). +Our [moderation cookbook](https://colab.research.google.com/github/mistralai/cookbook/tree/main/mistral/moderation/system-level-guardrails.ipynb) provides a concrete example of how to use the Moderation service to implement system level guardrails. +For a more broad view, there is also a more [explorative cookbook](https://colab.research.google.com/github/mistralai/cookbook/tree/main/mistral/moderation/moderation-explored.ipynb). ### FAQ Q: What is the distribution of false-positive and false-negative results on the new moderation API models. Specifically, will they be more likely to flag something as harmful when it is not or not flag something that is harmful? From 15a557d51cc62741cde5c12699dd9a5f511bbd50 Mon Sep 17 00:00:00 2001 From: pandora <128635000+pandora-s-git@users.noreply.github.com> Date: Wed, 4 Dec 2024 15:35:27 +0100 Subject: [PATCH 2/2] Update guardrailing.mdx --- docs/capabilities/guardrailing.mdx | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/capabilities/guardrailing.mdx b/docs/capabilities/guardrailing.mdx index 527902b..8faa777 100644 --- a/docs/capabilities/guardrailing.mdx +++ b/docs/capabilities/guardrailing.mdx @@ -183,8 +183,8 @@ The table below describes the types of content that can be detected in the moder ### Cookbook -Our [moderation cookbook](https://colab.research.google.com/github/mistralai/cookbook/tree/main/mistral/moderation/system-level-guardrails.ipynb) provides a concrete example of how to use the Moderation service to implement system level guardrails. -For a more broad view, there is also a more [explorative cookbook](https://colab.research.google.com/github/mistralai/cookbook/tree/main/mistral/moderation/moderation-explored.ipynb). +Our [moderation cookbook](https://colab.research.google.com/github/mistralai/cookbook/blob/main/mistral/moderation/system-level-guardrails.ipynb) provides a concrete example of how to use the Moderation service to implement system level guardrails. +For a more broad view, there is also a more [explorative cookbook](https://colab.research.google.com/github/mistralai/cookbook/blob/main/mistral/moderation/moderation-explored.ipynb). ### FAQ Q: What is the distribution of false-positive and false-negative results on the new moderation API models. Specifically, will they be more likely to flag something as harmful when it is not or not flag something that is harmful?