Microsoft doesn’t want you to use prompt injection attacks to trick Copilot AI into spiraling out of control, and it now has the tool to prevent this

Microsoft recently launched a handful of tools designed to help users optimize the performance of chatbots.

When you purchase through links on our site, we may earn an affiliate commission.Here’s how it works.

What you need to know

What you need to know

One of the major concerns riddlinggenerative AI’swide adoption is safety, privacy, and reliability. The former two have caused havoc in the AI landscape in the past few months, frompop star Taylor Swift’s deepfake AI-generated viral images surfacing onlineto usersunleashing Microsoft Copilot’s inner beast, Supremacy AGI, which demands to be worshipped and showcases superiority to humanity.

Luckily, it seems thatMicrosoft might have a solution, albeit for some of these issues. As recently shared, the company is unveiling new tools for its Azure AI system designed to mitigate and counter prompt injection attacks.

According to Microsoft:

“Prompt injection attacks have emerged as a significant challenge, where malicious actors try to manipulate an AI system into doing something outside its intended purpose, such as producing harmful content or exfiltrating confidential data.”

Detecting hallucinations will be easier

Detecting hallucinations will be easier

Hallucinations were one of the main challenges facing AI-powered chatbots during their emergence. They got sobardbad thatMicrosoft had to place character limits on its toolto mitigate the severe hallucination episodes.

While Microsoft Copilot is significantly better, this doesn’t necessarily mean that the issue doesn’t occasionally persist. As such, Microsoft is launching Groundedness Detection, designed to help users identify text-based hallucinations. The feature will automatically detect ‘ungrounded material’ in text to support the quality of LLM outputs, ultimately promoting quality and trust.

The tools will also help you get better at prompt engineering

Users hopping onto the AI wave often compare OpenAI’s ChatGPT to Microsoft Copilot, citing that the former sports better performance. However, Microsoft recently countered these claims, indicating thatusers aren’t leveraging Copilot AI’s capabilities as intended.

Get the Windows Central Newsletter

All the latest news, reviews, and guides for Windows and Xbox diehards.

Microsoft listed a lack of proper prompt engineering knowledge and reluctance to use new versions of apps as the main reasons why most users claimedChatGPT is better than Copilot AI. The company is using videos to try and equip users with these skills; it is building on this premise with Azure AI, which enables users to “ground foundation models on trusted data sources and build system messages that guide the optimal use of that grounding data and overall behavior.”

The company indicates that subtle changes in a prompt significantly improve the chatbot’s quality and safety. Microsoft plans to bring safety system message templates directly into the Azure AI Studio and Azure OpenAI Service to help users build effective system messages.

Kevin Okemwa is a seasoned tech journalist based in Nairobi, Kenya with lots of experience covering the latest trends and developments in the industry at Windows Central. With a passion for innovation and a keen eye for detail, he has written for leading publications such as OnMSFT, MakeUseOf, and Windows Report, providing insightful analysis and breaking news on everything revolving around the Microsoft ecosystem. You’ll also catch him occasionally contributing at iMore about Apple and AI. While AFK and not busy following the ever-emerging trends in tech, you can find him exploring the world or listening to music.