Beyond OpenAI: Microsoft Copilot flirts with Claude.

Microsoft has started to broaden their AI horizons by adding their first (not Open AI) model into Copilot.

Microsoft are integrating Anthropic’s Claude models into Microsoft 365 Copilot which marks a significant pivot from their exclusive OpenAI-centric approach. Microsoft are also working on their models which we already see on Copilot Plus PCs which will at some point make their way to Copilot.

This move is more than just a new menu option or toggle, it is part of their strategic play to diversify AI capabilities and reduce dependency on a single vendor.

Claude Opus and Sonnet in Copilot.

Claude Opus 4.1 and Sonnet 4 are now available to commercial Frontier Copilot users, (Corporate early adoptors) offering, for the first time, alternatives to Open AI’s GPT models for agents in Copilot Studio and also for their Researcher Agent.

Copilot Studio Model Selector (preview)

It’s worth noting that enabling access does require admin approval. See later.

In the formal annoucement, Microsoft said that Anthropic’s models will unlock “more powerful experiences” for users.

Claude Model option in Copilot Researcher.

Claude is not new to Microsoft however, with it already embedded in Visual Studio and Azure AI alongside Google’s AI and Elon Musk’s Grok. This is, however the first time Copilot launch that we have seen non OpenAI models powering Copilot.

Why This Matters

Microsoft’s shift to leveraging different models reflects a broader trend. Microsoft’s message here is that Copilot is no longer about a single model or even vendor, bit more about orchestration, choice, and adaptability.

Different models have different areas of excellence and this sets the foundations for Microsoft to give flexibilityto tailor and tune AI experiences to specific business needs  using the most appropriate model for the task.

It does, however, raise questions around governance, model performance, and cost. With multiple models in play, we don’t really know how the future of pricing will work if multi model is the future for Microsoft 365 Copilot.

Data Sovereignty and Multi-Model Concerns?

One question I’m already seeing is around Microsoft’s boundary of trust and responsibility, something Microsoft boast around with their Microsoft 365 portfolio.

While the flexibility of multi-model AI is compelling, the question is does it introduce new considerations around data residency and compliance when multi models are being used?

To address that, Microsoft has confirmed that these Claude models run within its Azure AI infrastructure, but states that are not Microsoft-owned. This means that when users “opt to” use Claude, their prompts and responses may be processed by Anthropic’s models hosted within Microsoft’s environment.

This means that when organisations choose to use Anthropic models, they are using these under Anthropic’s Commercial Terms of Service, not the consumer user terms.

For regulated industries or organisations with strict data governance policies, this is likely to raises a few red flags or at least questions that Microsoft will need to be able to answer.

  • Data Boundary Clarity: Is the data staying within Microsoft’s compliance boundary, or is it crossing into Anthropic’s operational domain? If so what does this mean for data compliance and security?
  • Model-Specific Logging: Are logs and telemetry handled differently across models? Can organisations audit usage per model? How is encrypted data handled?
  • Privacy and Consent: Are users aware when their data is being processed by a non-Microsoft model? Is consent granular enough? Will users understand even if Microsoft tell them?

Again, Microsoft has stated that Claude models are “fully integrated” into the Microsoft 365 compliance framework, but organisations will still want to (and should) validate this against their own risk posture – especially where sensitive or regulated data is involved.

Enabling Claude models in Copilot.

To enable the models, your Microsoft 365 Admin needs to head over to the Microsoft 365 Admin Centre and enable access to the other models. Instructions for this are shown in the link below.

https://learn.microsoft.com/en-us/copilot/microsoft-365/connect-to-ai-models?s=09

Thoughts.

This is a smart move I think. Microsoft is playing the long game — moving their eggs out of one basket and looking a different models that made most economic and performance sense and brining more choice to agent builders.

For those of us partners like us at Cisilion, advising clients on AI adoption, this reinforces the need to think modularly. When building agents, don’t just pick a model – pick a framework that allows you to evolve. Microsoft’s Copilot is becoming that framework and that should be good for business.

I do expect this is just the start. We know Microsoft’s relationship with OpenAI is “less properpous” that it once was. As such I do expect more models, more integrations, and more choice and I do think we will see Microsoft’s own models making their way to Copilot soon.

But with choice comes complexity. We need to ensure that governance, transparency, and user education keep pace with innovation. Again partners will need to help customers navigate this.

What do you think. Is this a good move for Microsoft and their customers?

Will OpenAI’s “gpt-realtime” set a new benchmark for AI Voice?

OpenAI has introduced gpt-realtime, a new cutting-edge speech-to-speech model, alongside the general availability of its Realtime API. This release marks a significant step forward in the evolution of voice AI, particularly for enterprise applications such as customer support and conversational agents. They announced this in a video broadcast which you watch below.

SIP Telephony Support: Lowering the Barrier to Entry

One of the most notable updates they annunced was the addition of SIP telephony support, which aims to simplify the process of building voice-over-phone applications. Developers will be able to integrate phone numbers directly into OpenAI’s SIP interface, streamlining deployment and reducing the need for complex telephony infrastructure. As it develops, this could reshape the competitive landscape, especially for startups that previously relied on expensive and bespoke integrations to differentiate their offerings.

A Unified Model for Natural Interaction

Gpt-realtime apart will feature an end-to-end architecture which will set it apart to how such integrations work today. Unlike traditional systems that chain together speech recognition, language processing, and text-to-speech, OpenAI’s new model will handle everything in a single pass. This will result in much faster response times, more natural audio, and improved emotional nuance (one of the biggest limitations today) meaning it will be capable of interpreting laughter, stress, worry, pauses, and tone shifts.

Open AI so it will also be highly configurable. Developers will be able to adjust pacing, tone, and persona, enabling more tailored and brand-consistent voice experiences.

Considerations for Enterprise Adoption

While the capabilities are lok super impressive, these models will still be expensive to start with anyway. Pricing is expected to be $32 per million input tokens and $64 per million output tokens which is significantly higher than traditional chained models. Additionally, the unified architecture offers less modularity and observability, which may limit flexibility for teams that require fine-grained control over model behavior or voice switching.

In a blog post from CX Today, Alex Levin, CEO at Regal is quoted saying,the cost of the speech-to-speech model is still approximately four times higher than chaining a speech-to-text (STT), large language model (LLM), text-to-speech (TTS) pipeline for Voice AI Agents

Strategic Implications

OpenAI’s latest release is a clear signal of intent: to make voice AI more accessible, performant, and enterprise-ready. Given Mcirosoft and other leading Cloud giants, close relationships with Open AI, we can also expect them to eventally add support for such models meaning customers that leverage, for example Microsoft 365 Copilot and Azure AI will likely gain support for this in the near future too through tools like Microsoft Dynamnics and Copilot Studio.

For organisations exploring and wanting to experiment more with conversational based automation, gpt-realtime promises to offers a powerful new toolset whilst talking the technolgy closer to human voice.

As always, the key lies in aligning technology choices with business goals, recognising ROI and customer expectation and keeping ahead of the curve as the landscape evolves and the pace of AI maturity and adoption contines to accelerate.


Sources: 
CX Today – OpenAI’s Latest Moves Put Many Voice AI Startups on Notice
Open AI – YouTube Video:
Open AI Blog

Microsoft makes OpenAI o1 model free for Copilot users.

OpenAI’s most advanced AI model “o1” which is known for its problem solving and deeper thinking has been available behind a $20 per month ChatGPT premium subscription. ChatGPT premium gives limited acess for $20 a month and unlimited access for $200 a month.

Copilot let’s you use it for free.

Microsoft has a tight partnership with OpenAI and is also on a mission to put their AI (Copilot) across every Microsoft Service it offers with huge capability and features even on theor “free” tiers.

Copilot Consumer Pro users have had access to Think Deeper (which uses the o1 model) for the past 12 months, but Microsoft have now made this feature free to everyone including those using the free version of Copilot.

To access it, you need to simply head ovee to Copilot on the web, (or via the mobile app) and ensure you are signed in with a Microsoft account (MSA). You then get completed free access to the Think Deeper search (which uses the o1 model).

How to get Microsoft Copilot

To get Copilot, head to the web (you actually find Copilot in the Edge browser) and go to https://copilot.microsoft.com or head over to you phones app store and search for Copilot and install it.

You need to be signed in with your Microsoft account to use these features.

Using o1 features aka Think Deeper

Once in Copilot, use the AI chat as you would before (or like you did in ChatGPT) and you will see a “think deeper” button inside the text input box.

Using Copilot’s Think Deeper (ChatGPT model o1)

Selecting it activates the o1 reasoning model. As it processed your prompt, you also get a spinning symbol since searches and responses using o1 are more thorough that with GPT 4 and typpically take around 30 secs.

Using Copilot’s Think Deeper.

This is Microsoft’s way of letting you know that you’re in for around a 20-30 seconds wait. If you don’t need deep search so for normal use), toggle this back off to go back to the super fast GPT-4o version…

So what can o1 do then?

The deep thinker feature of Microsoft Copilot is much better for more complex tasks and research due to the o1 model ability for in depth reasoning. 

As such it is simply better for solving complex issues like math, logic or science, for analysing or creating long or richer documents and reports or for code creation and debug. The best way to test this is to run two Copilot Windows side by side and test out the same prompt with and without Think Deeper enabled.

Content created with o1 is also more “accurate” with far less AI hallucinations (aka, making things up).

Why do many GPTs Hallucinate? In general, GPT models learn by mimicking patterns in their training data (huge amounts of data). The o1 model uses a different technique called reinforcement learning, whereby it's language model works things out (though it's training) by rewarding the right answers and penalising wrong ones. This takes longer through the iterative and testing process. Once done the model  moves through queries in a step-by-step fashion much like human problem  solving. 

o1 limitations?

It is worth noting that o1 isn’t quite on the same level as ChatGPT in some areas. It is less effective with factual knowledge and is currently less able to search the Internet and cannot process files and images.

What about DeepSeek?

The big story this week has of course been DeepSeek, a controversial Chinese AI firm that has announced and launched their own GPT-4 and o1 rivals that have been supposedly built at a fraction of the cost of OpenAI, Google and other US models, shaking share prices, disrupting the market and rasing many questions.

What is more is more is that DeepSeek models are claimed to be more advanced and faster than GPT-4o and smarter that o1.

The advent of DeepSeek has sent shockwaves through the tech industry. Global stock markets have reeled, sparking a cascade of investigations and looming threats of bans.

Yet, the bot hasn’t been without its champions. Interestly, Microsoft – OpenAI’s top financial invester and partner  – has already embraced the DeepSeek R1 reasoning model, and has integrating it into Azure AI Foundry and also GitHub.

These platforms, beloved by developers for fostering advanced AI projects, now stand as the new playground for DeepSeek’s innovative potential.

DeepSeek logo

Open AI Strikes Back

In the wake of its free mobile app’s viral triumph, OpenAI’s CEO Sam Altman swiftly revealed plans to accelerate the rollout of new releases to keep ahead of its new Chinese competitor.

OpenAI are not standing still either though. Et the end of December 2024, month, they began  trialing twin AI models, o3 and o3 mini. Remarkably, the former has surpassed o1 in coding, mathematics, and scientific capabilities, marking a significant advancement in their AI prowess.

There is no doubt this is an area that doesn’t stand still. By the time I click publish this post will likely already be out of date!


DeepSeek has certainly ignited an even greater sense of urgency within the already dynamic AI sector which moves and evolves on an almost daily basis.

Hands-on with Copilot Voice: An almost human conversation

Copilot Voice Cover


The recent Copilot update is a Game-Changer in AI Voice Technology. In the recent announcements, Microsoft unveiled a new version of its Copilot app for iPhone and Android. The update brings a fresh look and new features. It also includes an impressive voice mode that rivals OpenAI’s ChatGPT Advanced Voice – especially since Microsoft make this available for free – yes free

I have tested both recently. I can confidently say that the new Copilot is a significant upgrade. What’s more, it is totally free to use. This is best read while/after you have watched my hands-on video below.

Hands on with Copilot Voice

User-Friendly Interface and Enhanced Voice Mode

The updated Copilot app boasts a more “consumer-friendly” interface. I do wish they would bring some of the advanced customisations back. The standout feature in this update is most definitely the new voice mode, which on first look (a few app updates before it worked), I thought would be a bit of a fad – but it is absolutely brilliant.

Voice mode offers speech-to-speech functionality, allowing for more natural and engaging conversations. While it may not interrupt as fluidly as OpenAI’s offering (though it’s still in early stages), it feels more casual and less stilted, making interactions feel more like chatting with a friend.

A Conversation That Feels Real

During my testing, I found myself deeply starting to actually forget that I was talking to an AI as the conversation felt natural and real (there was the odd delay. In my hands-on example (see the video below), I participated in a discussion. We talked about “if and when AI could ever become self-aware”. We also considered what the implications might be. Unlike a text-based discussion, this level of engagement goes to show just how fast and how rapid the advancement of natural conversation is becoming.

Copilot appears to adapt its vocal tones and pace during conversations. It emphasizes certain words as we speak.

Perhaps the biggest (pleasant) surprise I found was how Copilot adapted to use slang terms the more I used them too. If I swore or spoke more loudly, it also seemed to detect the change in my tone and adjust its output. I’ll be testing this more to see just how far it can go.

Spoiler: I did find the occasional limitation as the conversation continued, such as occasional delays when I interrupted it and seconds of silence.

Customisation and Accessibility

Copilot offers four voice options: Grove, Canyon, Wave, and Meadow. Unlike ChatGPT, you can modify the speed and tone of these voices, making them sound more natural and suited to your preferences. This feature, combined with the app’s inclination to use slang and short-hand words, makes it easy to forget you’re interacting with a machine. I’m not a fan of all the voices though and they are not currently that localised – with most very American (which is fine for now).

Gemini Live (yes, all the chat bots are discovering their voice) currently gives users a choice of 10, but Microsoft say more voice options will be coming “soon”.

What I also like is that you can customise the speed at which each of the voices speaks. Personally, I find the standard setting is too slow and find that a speed of 1.1x sounds most natural. I also discovered that you can also ask Copilot to speak differently by explaining how you want it to sound – for example, applying a slightly different accent, changing its tone of voice or to be more empathetic but I’d like to think eventually Copilot will do this natively without me asking (after all it’s unlikely you’d ask a human to speak in a different tone!).

Copilot Voice is free

One of the most significant advantages of Copilot is that it’s free to use. Today, OpenAI’s ChatGPT Advanced Voice feature currently requires a $20 monthly subscription, whilst Microsoft makes this feature available to all Copilot users, regardless of their subscription status.

Conclusion

Copilot is now under the leadership of Mustafa Suleyman [Microsoft CEO for AI]. It seems poised to make a significant impact in the AI voice technology market. It builds on its partner OpenAI. Its user-friendly design, natural voice interactions, and accessibility make it a strong competitor against other AI voice models.

The best thing – this is totally free

Try this out. Let me know how in depth you feel during a conversation is and can be with Copilot. How “close” do you think this is in becoming a natural, almost human conversation.

Microsoft confirm GPT-4o is now available on Azure AI

Just ahead of Microsoft Build, the Azure team have announced the availability of GPT-4o, OpenAI’s latest flagship model on Azure AI. This innovative multimodal model combines text, vision, and audio capabilities, establishing a new benchmark for generative and conversational AI experiences. GPT-4o is now available in the Azure OpenAI Service for preview, with support for text and image inputs.

This is a preview for testing now

What does GPT-4o Bring?

GPT-4o represents a paradigm shift in the interaction of AI models with multimodal inputs. It integrates text, images, and audio to deliver a more immersive and engaging user experience.

What does the “preview” include?

Currently in preview, Azure OpenAI Service customers will be able to test GPT-4o’s broad capabilities via a preview playground in Azure OpenAI Studio. This initial version emphasizes text and visual inputs, offering a preview of the model’s possibilities and setting the stage for additional functionalities, including audio and video.

The preview is free to try but has limitations around usage and location availability.

Designed for rapidity and efficiency, GPT-4o’s sophisticated processing of complex inquiries with fewer resources has the potential to offer both cost efficiency and enhanced performance.

Note: At time of writing, this is preview is available in two US regions only West US3 and East US.

What about GPT-4o in Microsoft Copilot?

We don’t know yet, but we do know that there will exciting updates around the rest of the Microsoft AI stack this week. Microsoft has an agressive and innovation fuelled roadmap for Microsoft 365 Copilot so as Microsoft continues to update and integrate OpenAI’s latest models into Copilot – I’m looking forward to hearing more this week.

What else is coming?

This week is Microsoft Build 2024 in Seattle and online. I expect this to be (pretty much) all about Copilot, and AI so expect to hear more about GPT-4o and other Azure AI updates.


Further Reading

You can read more about GOT-4o at the official OpenAI Blog which is < here >.

Microsoft are adding a Copilot for Copilot (well sort of).

Yesterday, (8th May, 24) Microsoft released their 2024 Work Trend Index Report which covered the State of AI at Work (you can see this here) as well as announcing some more improvements coming to Copilot for Microsoft 365 in the coming months.

The new features annouced are all aimed at helping to optimise prompt writing, making it easier for people to get a prompt that does what they need first time (a Copilot for Copilot essentially). These updates will include.

  • Auto-complete for prompts
  • Prompt re-write
  • A new catch up feature
  • Copilot Labs upgrade.

Let dive into these quickly. All. Images (c) Microsoft.

Auto Complete for Prompts

Copilot’s new “autocomplete” feature is similar to what you get in a search engine, where it will anticipate (using Machine Learning) what you are writing and help you to complete your prompt when you start typing one out.

Image (c) Microsoft

The aim here to suggesting more details to ensure you get the intended outcome. It will also offer an expanded library of ‘next prompts’.

This means if you start typing “summarise” then Copilot will display options to summarise the last 10 unread emails and chat messages or other tasks that might be related.

Prompt Rewrite

The “rewrite” feature is something that many image AI tools have had for a while. The aim is to be able to takes a person’s basic prompt rewrite it to me more thorough, “turning everyone into a prompt engineer,” according to the Microsoft.

Image (c) Microsoft

Also known as “elaborate your prompt”, Microsoft say this will be able to rewrite any prompts people create making it much easier to do more complex tasks especially when working with documents or ‘connected apps’.

Copilot Catch-up

Copilot Catch Up aims to start making Copilot more “proactive”. Here the chat interface will be able to presents people with “responsive recommendations” based on their recent activity. As an example, it will be able to notify you about upcoming meetings and suggest ways to help you prepare for that meeting, by bringing a summary of recent email and chat threads, meetings notes and documents write in the chat thread. This feature is also coming into Copilot in Outlook.

This feature brings Copilot more into the realms of good ol Clippy (ok I’m kidding here) but will enable Copilot to start proactively helping rather than waiting for its pilot to issue a command and bring the genie out of its lamp!

The aim is to further integrate Copilot into the user’s workflows. Imagine for example having a morning prompt that tells you about your day, tickets logged via Service Now, or a project that is over running (via Project or Planner) or has  completed early perhaps!

Updates to Copilot Labs

Similar to Microsoft app Prompt Buddy, Microsoft will also start to allow people to create, publish, and manage prompts in Copilot Lab.

Image (c) Microsoft

This will bring new features that that can be tailored for individual teams within businesses. This is aimed to make it a lot easier to share useful prompts for employees, Teams and departments to use.

Will these help adoption?

What do you think about the new updates, will these help remove the dark art of promoting and make Copilot easier to use and faster to help people get the desired results.?

Let me know on the comments..

Microsoft to open new AI Hub in London

Microsoft has announced plans for a new artificial intelligence (AI) hub in London, which will be focused on leading edge product development and research. This will be led Microsoft AI Lead Mustafa Suleyman (confounder of DeepMind) who Microsoft hired last month.

This annoucement comes less than a month since Microsoft unveiled a new consumer AI division.

There is an enormous pool of AI talent and expertise in the UK, and Microsoft AI plans to make a significant, long-term investment in the region. (London).

Mustafa Suleyman

This is great for the UK and for London and will help both Microsoft and the UK become an AI  and technology superpower leveraging the hub of tech talent, access to leading and world class universities and research centres with ability to attract the best talent for the next generation of development of AI.

Microsofts AI Future in the UK

This announcement builds on Microsoft’s recent commitment to invest 2.5 Billion into data centre infrastructure and improving AI skills across the UK.

Microsoft’s AI investment in the UK includes building a major new data centre in West London and installing 20,000 high-powered processors in the UK by 2026.

Microsoft’s new UK hub will be run by Jordan Hoffmann,  (another former employee from DeepMind) and will collaborate closely with OpenAI which powers Microsoft’s AI driven Copilot System framework.