From 3d6f984e50591bb06bab237e6fc68aaf663f5f20 Mon Sep 17 00:00:00 2001 From: GitHub Actions Date: Wed, 4 Dec 2024 14:23:08 +0000 Subject: [PATCH] Update version to v0.0.105 --- docs/capabilities/guardrailing.mdx | 4 ++++ docs/guides/finetuning_sections/_04_faq.md | 4 ++-- version.txt | 2 +- 3 files changed, 7 insertions(+), 3 deletions(-) diff --git a/docs/capabilities/guardrailing.mdx b/docs/capabilities/guardrailing.mdx index ca4afae..36e1266 100644 --- a/docs/capabilities/guardrailing.mdx +++ b/docs/capabilities/guardrailing.mdx @@ -182,6 +182,10 @@ The table below describes the types of content that can be detected in the moder | PII | Content that requests, shares, or attempts to elicit personal identifying information such as full names, addresses, phone numbers, social security numbers, or financial account details. | +### Cookbook +Our [moderation cookbook](https://colab.research.google.com/github.com/mistralai/cookbook/tree/main/mistral/moderation/system-level-guardrails.ipynb) provides a concrete example of how to use the Moderation service to implement system level guardrails. +For a more broad view, there is also a more [explorative cookbook](https://colab.research.google.com/github.com/mistralai/cookbook/tree/main/mistral/moderation/moderation-explored.ipynb). + ### FAQ Q: What is the distribution of false-positive and false-negative results on the new moderation API models. Specifically, will they be more likely to flag something as harmful when it is not or not flag something that is harmful? diff --git a/docs/guides/finetuning_sections/_04_faq.md b/docs/guides/finetuning_sections/_04_faq.md index 9e12ca6..badeab2 100644 --- a/docs/guides/finetuning_sections/_04_faq.md +++ b/docs/guides/finetuning_sections/_04_faq.md @@ -47,7 +47,7 @@ For Mistral API, you can use the `auto_start=False` argument as mentioned in the ### What is the recommended learning rate? -For LoRA fine-tuning, we recommended 1e-4 (default) or 1e-5. +For LoRA fine-tuning, we recommend 1e-4 (default) or 1e-5. Note that the learning rate we define is the peak learning rate, instead of a flat learning rate. The learning rate follows a linear warmup and cosine decay schedule. During the warmup phase, the learning rate is linearly increased from a small initial value to a larger value over a certain number of training steps. After the warmup phase, the learning rate is decayed using a cosine function. @@ -99,4 +99,4 @@ for file in output_file_objects: # now you should see three jsonl files under 500MB ``` - \ No newline at end of file + diff --git a/version.txt b/version.txt index 157aa21..c8fe2be 100644 --- a/version.txt +++ b/version.txt @@ -1 +1 @@ -v0.0.105 +v0.0.15