|
| 1 | +# Content Filtering \& Guardrails - Overview |
| 2 | + |
| 3 | +Costa Rica |
| 4 | + |
| 5 | +[](https://github.com/) |
| 6 | +[brown9804](https://github.com/brown9804) |
| 7 | + |
| 8 | +Last updated: 2025-12-26 |
| 9 | + |
| 10 | +------------------------------------------ |
| 11 | + |
| 12 | +> Microsoft Foundry enforces Responsible AI principles by applying content filters to all Large Language Models (LLMs) and image generation models. |
| 13 | +
|
| 14 | +<details> |
| 15 | +<summary><b>List of References</b> (Click to expand)</summary> |
| 16 | + |
| 17 | +- [Content filtering overview](https://learn.microsoft.com/en-us/azure/ai-foundry/openai/concepts/content-filter?view=foundry-classic) |
| 18 | +- [Configure content filters](https://learn.microsoft.com/en-us/azure/ai-foundry/openai/how-to/content-filters?view=foundry-classic#understand-content-filter-configurability) |
| 19 | +- [Content filtering in Microsoft Foundry portal](https://learn.microsoft.com/en-us/azure/ai-foundry/concepts/content-filtering?view=foundry-classic) |
| 20 | +- [Azure OpenAI in Microsoft Foundry model deprecations and retirements](https://learn.microsoft.com/en-us/azure/ai-foundry/openai/concepts/model-retirements?view=foundry-classic&tabs=text#current-models) |
| 21 | +- [Emissions Impact Dashboard for Azure](https://marketplace.microsoft.com/en-us/product/power-bi/coi-sustainability.emissions_impact_dashboard) |
| 22 | +- [GPT‑5.1 in Foundry: A Workhorse for Reasoning, Coding, and Chat](https://techcommunity.microsoft.com/blog/partnernews/gpt%E2%80%915-1-in-foundry-a-workhorse-for-reasoning-coding-and-chat/4469803) |
| 23 | +- [Getting started with Azure OpenAI Assistants (Preview)](https://learn.microsoft.com/en-us/azure/ai-foundry/openai/how-to/assistant?view=foundry-classic) |
| 24 | + |
| 25 | +</details> |
| 26 | + |
| 27 | +> [!NOTE] |
| 28 | +> These filters are powered by Azure AI Content Safety and operate at two stages: |
| 29 | +> - Input filtering: Checks user prompts before sending them to the model. |
| 30 | +> - Output filtering: Checks model responses before returning them to the user. |
| 31 | +
|
| 32 | +> Default Filters: |
| 33 | +> - Every deployment has a `default filter applied automatically.` |
| 34 | +> - These filters `detect harmful content in four categories:` |
| 35 | +> - Violence |
| 36 | +> - Hate |
| 37 | +> - Sexual |
| 38 | +> - Self-harm |
| 39 | +> - Each category has severity levels: safe, low, medium, high. |
| 40 | +> - By default, medium and high severity content is blocked, while safe and low are allowed. |
| 41 | +
|
| 42 | +## Custom Filters |
| 43 | + |
| 44 | +> You can create custom content filters to override defaults: |
| 45 | +> - Configure severity thresholds per category. |
| 46 | +> - Enable or disable binary classifiers (e.g., jailbreak detection). |
| 47 | +> - Custom filters can be applied at resource level or deployment level using: |
| 48 | +> - Foundry portal |
| 49 | +> - Azure CLI |
| 50 | +> - Bicep templates |
| 51 | +
|
| 52 | +> [!IMPORTANT] |
| 53 | +> By default, user's, any organizational data that flows through those agents is not shared publicly and is not used to train other models. |
| 54 | +
|
| 55 | +<!-- START BADGE --> |
| 56 | +<div align="center"> |
| 57 | + <img src="https://img.shields.io/badge/Total%20views-1633-limegreen" alt="Total views"> |
| 58 | + <p>Refresh Date: 2025-12-03</p> |
| 59 | +</div> |
| 60 | +<!-- END BADGE --> |
0 commit comments