Skip to content

Commit dda9521

Browse files
authored
docs: prepare 0.18 doc release (#1493)
* prepare 0.18 doc release * add the other two notes * nit
1 parent b5b7579 commit dda9521

File tree

4 files changed

+34
-1
lines changed

4 files changed

+34
-1
lines changed

docs/project.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
{ "name": "nemo-guardrails-toolkit", "version": "0.17.0" }
1+
{ "name": "nemo-guardrails-toolkit", "version": "0.18.0" }

docs/release-notes.md

Lines changed: 27 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,33 @@ The following sections summarize and highlight the changes for each release.
1212
For a complete record of changes in a release, refer to the
1313
[CHANGELOG.md](https://github.com/NVIDIA/NeMo-Guardrails/blob/develop/CHANGELOG.md) in the GitHub repository.
1414

15+
---
16+
17+
(v0-18-0)=
18+
19+
## 0.18.0
20+
21+
(v0-18-0-features)=
22+
23+
### Key Features
24+
25+
- In-memory caching of guardrail model calls for reduced latency and cost savings.
26+
NeMo Guardrails now supports per-model caching of guardrail responses using an LFU (Least Frequently Used) cache.
27+
This feature is particularly effective for safety models such as NVIDIA NemoGuard [Content Safety](https://build.nvidia.com/nvidia/llama-3_1-nemoguard-8b-content-safety), [Topic Control](https://build.nvidia.com/nvidia/llama-3_1-nemoguard-8b-topic-control), and [Jailbreak Detection](https://build.nvidia.com/nvidia/nemoguard-jailbreak-detect) where identical inputs are common.
28+
For more information, refer to [](model-memory-cache).
29+
- NeMo Guardrails extracts the reasoning traces from the LLM response and emits them as `BotThinking` events before the final `BotMessage` event.
30+
For more information, refer to [](bot-thinking-guardrails).
31+
- New community integration with [Cisco AI Defense](https://www.cisco.com/site/ca/en/products/security/ai-defense/index.html).
32+
- New embedding integrations with Azure OpenAI, Google, and Cohere.
33+
34+
(v0-18-0-fixed-issues)=
35+
36+
### Fixed Issues
37+
38+
- Implemented validation of content safety and topic control guardrail configurations at creation time, providing prompt error reporting if required prompt templates or parameters are missing.
39+
40+
---
41+
1542
(v0-17-0)=
1643

1744
## 0.17.0

docs/user-guides/advanced/bot-thinking-guardrails.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,5 @@
1+
(bot-thinking-guardrails)=
2+
13
# Guardrailing Bot Reasoning Content
24

35
Reasoning-capable large language models (LLMs) expose their internal thought process as reasoning traces. These traces reveal how the model arrives at its conclusions, providing transparency into the decision-making process. However, they may also contain sensitive information or problematic reasoning patterns that need to be monitored and controlled.

docs/versions1.json

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,10 @@
11
[
22
{
33
"preferred": true,
4+
"version": "0.18.0",
5+
"url": "../0.18.0/"
6+
},
7+
{
48
"version": "0.17.0",
59
"url": "../0.17.0/"
610
},

0 commit comments

Comments
 (0)