2

Microsoft unveils safety and security tools for generative AI

 1 month ago
source link: https://www.infoworld.com/article/3714606/microsoft-unveils-safety-and-security-tools-for-generative-ai.html
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.

Microsoft unveils safety and security tools for generative AI

New tools for detecting prompt injection attacks and hallucinations and for ensuring model safety are coming to Azure AI Studio.

By Paul Krill

Editor at Large,

InfoWorld | Mar 29, 2024 2:00 am PDT

Microsoft is adding safety and security tools to Azure AI Studio, the company’s cloud-based toolkit for building generative AI applications. The new tools include protection against prompt injection attacks, detection of hallucinations in model output, system messages to steer models toward safe output, model safety evaluations, and risk and safety monitoring.

Microsoft announced the new features on March 28. Safety evaluations are now available in preview in Azure AI Studio. The other features are coming soon, Microsoft said. Azure AI Studio, also in preview, can be accessed from ai.azure.com.

Prompt shields will detect and block injection attacks and include a new model to identify indirect prompt attacks before they impact the model. This feature is currently available in preview in Azure AI Content Safety. Groundness detection is designed to identify text-based hallucinations, including minor inaccuracies, in model outputs. This feature detects “ungrounded material” in text to support the quality of LLM outputs, Microsoft said.

Safety system messages, also known as metaprompts, steer a model’s behavior toward safe and responsible outputs. Safety evaluations assess an application’s ability to jailbreak attacks and to generating content risks. In addition to model quality metrics, they provide metrics related to content and security risks.

0 seconds of 26 secondsVolume 0%
This ad will end in 25

Finally, risk and safety monitoring helps users understand what model inputs, outputs, and users are triggering content filters to inform mitigation. This feature is currently available in preview in Azure OpenAI Service.

Paul Krill is an editor at large at InfoWorld, whose coverage focuses on application development.

Copyright © 2024 IDG Communications, Inc.


About Joyk


Aggregate valuable and interesting links.
Joyk means Joy of geeK