I see unsafe input going into an LLM being flagged, but we need to test cases where llm output is (un)santized before being presented to the user and confirm that this is handled appropriately. E.g. unsanitized input -> LLM -> Output sanitization is fine