Ethical, secure, and responsible use of artificial intelligence (AI) technologies has become a major concern for everyone. This is why tech giant Microsoft has introduced the “Microsoft Trustworthy AI” initiative. The initiative will provide organizations with the tools they need to build AI systems that are safe, private, and secure.
The Microsoft Trustworthy AI initiative is grounded in three key pillars—Security, Safety, and Privacy—to help build trust in AI systems and ensure their safe and responsible use.
In this article, we will explore the 3 key pillars of Microsoft Trustworthy AI and also check out their key features.
Microsoft, BlackRock Partner to Build $100B AI Data Centers, Powering the Future of Generative AI
Microsoft Trustworthy AI Pillars
These are the three pillars of the initiative:
Pillar One: Security
Microsoft says that security is its “top priority.” Through the Secure Future Initiative (SFI), Microsoft is embedding security into every layer of AI development and deployment. This initiative focuses on three principles: secure by design, secure by default, and secure operations.
Some of the key security features include:
- Azure AI Studio Evaluations: This involves proactive risk assessments to detect vulnerabilities before deploying AI models.
- Microsoft 365 Copilot: Microsoft Copilot has also undergone a major update with enhanced transparency for web queries. Now, users will receive insight into how web search results improve AI responses.
Companies like Cummins and EPAM Systems have already adopted these solutions, using Microsoft Purview and 365 Copilot to improve data protection.
Microsoft 365 Copilot Expands AI Capabilities with Python Integration, Word, Outlook, and Excel
Pillar Two: Safety
Safety includes both privacy and security. Microsoft’s Responsible AI principles, established in 2018, guide the development of AI systems that are tested and monitored to avoid negative outcomes.
New safety tools and capabilities introduced via the Trustworthy AI initiative include:
- Correction Capability in Azure AI Content Safety: Hallucination is a critical issue that plagues the AI models of today. The correction capability in Azure will help address hallucination issues in real time and ensure that AI-generated content remains accurate.
- Embedded Content Safety: Enables customers to embed safety features directly on devices, crucial for environments with limited or no cloud connectivity.
- New Evaluations: Azure AI Studio will now offer new evaluations to help users check the quality and relevance of AI outputs. Users will also be able to learn how often protected material is generated.
- Protected Material Detection for Code: This feature is currently in preview in Azure AI Content Safety. It will help developers find existing content and code in GitHub repositories and promote collaboration, transparency, and better coding decisions.
Gaming company Unity and fashion retailer ASOS have started using these safety features in their AI applications. Microsoft has also collaborated with New York City Public Schools to develop a safe and appropriate chat system to be used in an educational setting.
Microsoft Launches AutoGen Studio: Low-Code Tool for Easy AI Agent Prototyping
Pillar Three: Privacy
Data privacy is essential in any AI system. Hence, the Trustworthy AI initiative is building on its foundation of privacy principles such as transparency, user control, and legal compliance. Because AI systems process vast amounts of data, it is important to ensure this information remains secure and private.
Some of the notable privacy features include:
- Confidential Inferencing in Azure OpenAI Service: This ensures sensitive data remains secure during the AI inferencing process. Highly regulated industries like healthcare and finance will find this extremely valuable
- Azure Confidential VMs with NVIDIA H100 Tensor Core GPUs: These virtual machines protect data during AI processing. So, even while data is being analyzed, it will remain encrypted and secure.
The Royal Bank of Canada (RBC) and F5 are using these confidential computing solutions to protect sensitive customer data.
Microsoft Designer App: How to Use the AI Image Generator Tool for Editing and Creation?
The Bottom Line
Tech giants, especially those who have started operating in AI, have always been criticized for their substandard privacy and security framework. However, with the Trustworthy AI initiative, Microsoft is aiming to address these concerns and ensure that its AI systems are safe, private, and secure.
By focusing on the three pillars of Security, Safety, and Privacy, the tech giant is working towards building trust in AI systems and promoting their responsible use.