Microsoft unleashes autonomous Copilot AI agents in public preview
They can learn, adapt, and make decisions – but don't worry, they're not coming for your job
Ignite Microsoft has fresh tools out designed to help businesses build software agents powered by foundation models – overenthusiastically referred to as artificial intelligence, or AI.
"Our vision is to empower every employee with a personal assistant via Copilot that allows them to tackle work's biggest pain points like meeting overload, overflowing inbox, designing sales decks in less time and with more business impact," comms chief Frank Shaw told media in a briefing.
Copilot, in case you've missed it, is the term Microsoft has adopted to refer to its AI productivity software. Based on OpenAI's GPT-4 series of large language models, it provides a chatbot interface through which users can enter an input prompt (text, audio, or an image, depending on context) and receive some response.
Microsoft 365 Copilot lets users tell Word to draft or edit text on whatever topic the chatbot can satisfactorily reproduce from its undisclosed training data, ask PowerPoint to create slides, or direct Excel to build a data visualization for a data set – among other things.
The next step for the technology is AI agents – chatbots that perform a series of linked tasks based on instructions. Microsoft teased its autonomous agent capabilities last month at an AI event in the UK, and they have graduated to public preview.
Agents represent a modern take on macros and workflow automation. The hope is that they can parrot human digital output well enough to pass inspection.
"Whether it's sourcing sales leads, handling customer service inquiries, or tracking project deadlines, agents can do a lot of the heavy lifting so people can focus on more strategic tasks," explained Shaw.
What separates agents from models, according to Microsoft Source writer Susanna Ray, is memory, entitlements, and tools. Memory lets agents perform a series of tasks that build on one another. Entitlements ensure the code has permission to access data and take action. And tools refer to the necessary glue code or applications.
To realize its vision of software-directed action, Microsoft is introducing Microsoft 365 Copilot agents with predefined roles. These include:
- Agents in SharePoint, which "can be edited to have a personalized name and certain behaviors, and can be shared across emails, meetings and chats, with users being able to ask the agents questions and getting real-time responses." (It remains to be seen whether naming agents makes users more reluctant to delete them.)
- The Employee Self-Service Agent in Microsoft 365 Copilot Business Chat (in private preview), which can respond to HR and IT questions, fetch benefits and payroll info, and request hardware from IT, among other policy-oriented tasks. Imagine a dynamic corporate policy site that can make API calls.
- The Facilitator agent (in preview), which can take notes in Teams and can summarize the conversations participants hopefully can follow on their own.
- The Interpreter agent (next year) promises real-time interpretation in Teams meetings in up to nine languages. Better still, maybe, "meeting participants can also have the Interpreter simulate their personal voice for a more inclusive experience." So, voice cloning, then? Zoom is pursuing something similar.
- The Project Manager agent, it's claimed, can automate project management, from planning to execution.
And if these templates fail to satisfy, Microsoft Copilot Studio provides a way to customize AI agent behavior.
- Letting chatbots run robots ends as badly as you'd expect
- Google reportedly developing an AI agent that can control your browser
- Voice-enabled AI agents can automate everything, even your phone scams
- AI agent promotes itself to sysadmin, trashes boot sequence
"Agents built in Copilot Studio can operate independently, dynamically planning and learning from processes, adapting to changing conditions, and making decisions without the need for constant human intervention," explained Charles Lamanna, corporate VP of business and industry for Copilot, in a blog post. "These autonomous agents can be triggered by data changes, events, and other background tasks – and not just through chat!"
Copilot Studio includes templates for common agent scenarios that can serve as the basis for a customized version. It is also gaining support for voice-enabled agents, image uploading (for analysis by GPT-4o), and knowledge tuning – the ability to add new sources of knowledge to help agents respond to questions.
Developers interested in creating agents can use the Agent SDK to access services from Azure AI, Semantic Kernel, and Copilot Studio. There's also an Azure AI Foundry integration that links Copilot Studio to facilitate connection to services like Azure AI Search and the Azure AI model catalog.
Separately, there's a public preview of agent builder in Power Apps – Microsoft's low-code service for business apps.
Sarah Bird, chief product officer for Responsible AI, noted in a blog post that extra safety considerations arise with autonomous agents – something said of robot agents too – and that Microsoft is focused on ensuring that they behave. She argued that the standard Responsible AI practices – including the Copilot Control System, intended to allow IT departments to manage data access for Copilot and AI agents – can mitigate risks.
Microsoft's post on the subject observes that many agents include a human-in-the-loop check to make sure autonomous decision-making doesn't go off the rails. Nothing demonstrates confidence in automation more than a manual approval bottleneck.
Those looking to get a sense of AI agents in the field may wish to consider the Hiring Assistant launched by Microsoft's LinkedIn subsidiary. The bot was deployed to lighten the burden on human resources professionals who have a hard time dealing with the deluge of job applications and other clerical chores – a situation exacerbated by automation and AI. ®