Hands-on with Copilot Voice: An almost human conversation

Copilot Voice Cover


The recent Copilot update is a Game-Changer in AI Voice Technology. In the recent announcements, Microsoft unveiled a new version of its Copilot app for iPhone and Android. The update brings a fresh look and new features. It also includes an impressive voice mode that rivals OpenAI’s ChatGPT Advanced Voice – especially since Microsoft make this available for free – yes free

I have tested both recently. I can confidently say that the new Copilot is a significant upgrade. What’s more, it is totally free to use. This is best read while/after you have watched my hands-on video below.

Hands on with Copilot Voice

User-Friendly Interface and Enhanced Voice Mode

The updated Copilot app boasts a more “consumer-friendly” interface. I do wish they would bring some of the advanced customisations back. The standout feature in this update is most definitely the new voice mode, which on first look (a few app updates before it worked), I thought would be a bit of a fad – but it is absolutely brilliant.

Voice mode offers speech-to-speech functionality, allowing for more natural and engaging conversations. While it may not interrupt as fluidly as OpenAI’s offering (though it’s still in early stages), it feels more casual and less stilted, making interactions feel more like chatting with a friend.

A Conversation That Feels Real

During my testing, I found myself deeply starting to actually forget that I was talking to an AI as the conversation felt natural and real (there was the odd delay. In my hands-on example (see the video below), I participated in a discussion. We talked about “if and when AI could ever become self-aware”. We also considered what the implications might be. Unlike a text-based discussion, this level of engagement goes to show just how fast and how rapid the advancement of natural conversation is becoming.

Copilot appears to adapt its vocal tones and pace during conversations. It emphasizes certain words as we speak.

Perhaps the biggest (pleasant) surprise I found was how Copilot adapted to use slang terms the more I used them too. If I swore or spoke more loudly, it also seemed to detect the change in my tone and adjust its output. I’ll be testing this more to see just how far it can go.

Spoiler: I did find the occasional limitation as the conversation continued, such as occasional delays when I interrupted it and seconds of silence.

Customisation and Accessibility

Copilot offers four voice options: Grove, Canyon, Wave, and Meadow. Unlike ChatGPT, you can modify the speed and tone of these voices, making them sound more natural and suited to your preferences. This feature, combined with the app’s inclination to use slang and short-hand words, makes it easy to forget you’re interacting with a machine. I’m not a fan of all the voices though and they are not currently that localised – with most very American (which is fine for now).

Gemini Live (yes, all the chat bots are discovering their voice) currently gives users a choice of 10, but Microsoft say more voice options will be coming “soon”.

What I also like is that you can customise the speed at which each of the voices speaks. Personally, I find the standard setting is too slow and find that a speed of 1.1x sounds most natural. I also discovered that you can also ask Copilot to speak differently by explaining how you want it to sound – for example, applying a slightly different accent, changing its tone of voice or to be more empathetic but I’d like to think eventually Copilot will do this natively without me asking (after all it’s unlikely you’d ask a human to speak in a different tone!).

Copilot Voice is free

One of the most significant advantages of Copilot is that it’s free to use. Today, OpenAI’s ChatGPT Advanced Voice feature currently requires a $20 monthly subscription, whilst Microsoft makes this feature available to all Copilot users, regardless of their subscription status.

Conclusion

Copilot is now under the leadership of Mustafa Suleyman [Microsoft CEO for AI]. It seems poised to make a significant impact in the AI voice technology market. It builds on its partner OpenAI. Its user-friendly design, natural voice interactions, and accessibility make it a strong competitor against other AI voice models.

The best thing – this is totally free

Try this out. Let me know how in depth you feel during a conversation is and can be with Copilot. How “close” do you think this is in becoming a natural, almost human conversation.

Microsoft confirm GPT-4o is now available on Azure AI

Just ahead of Microsoft Build, the Azure team have announced the availability of GPT-4o, OpenAI’s latest flagship model on Azure AI. This innovative multimodal model combines text, vision, and audio capabilities, establishing a new benchmark for generative and conversational AI experiences. GPT-4o is now available in the Azure OpenAI Service for preview, with support for text and image inputs.

This is a preview for testing now

What does GPT-4o Bring?

GPT-4o represents a paradigm shift in the interaction of AI models with multimodal inputs. It integrates text, images, and audio to deliver a more immersive and engaging user experience.

What does the “preview” include?

Currently in preview, Azure OpenAI Service customers will be able to test GPT-4o’s broad capabilities via a preview playground in Azure OpenAI Studio. This initial version emphasizes text and visual inputs, offering a preview of the model’s possibilities and setting the stage for additional functionalities, including audio and video.

The preview is free to try but has limitations around usage and location availability.

Designed for rapidity and efficiency, GPT-4o’s sophisticated processing of complex inquiries with fewer resources has the potential to offer both cost efficiency and enhanced performance.

Note: At time of writing, this is preview is available in two US regions only West US3 and East US.

What about GPT-4o in Microsoft Copilot?

We don’t know yet, but we do know that there will exciting updates around the rest of the Microsoft AI stack this week. Microsoft has an agressive and innovation fuelled roadmap for Microsoft 365 Copilot so as Microsoft continues to update and integrate OpenAI’s latest models into Copilot – I’m looking forward to hearing more this week.

What else is coming?

This week is Microsoft Build 2024 in Seattle and online. I expect this to be (pretty much) all about Copilot, and AI so expect to hear more about GPT-4o and other Azure AI updates.


Further Reading

You can read more about GOT-4o at the official OpenAI Blog which is < here >.

Microsoft are adding a Copilot for Copilot (well sort of).

Yesterday, (8th May, 24) Microsoft released their 2024 Work Trend Index Report which covered the State of AI at Work (you can see this here) as well as announcing some more improvements coming to Copilot for Microsoft 365 in the coming months.

The new features annouced are all aimed at helping to optimise prompt writing, making it easier for people to get a prompt that does what they need first time (a Copilot for Copilot essentially). These updates will include.

  • Auto-complete for prompts
  • Prompt re-write
  • A new catch up feature
  • Copilot Labs upgrade.

Let dive into these quickly. All. Images (c) Microsoft.

Auto Complete for Prompts

Copilot’s new “autocomplete” feature is similar to what you get in a search engine, where it will anticipate (using Machine Learning) what you are writing and help you to complete your prompt when you start typing one out.

Image (c) Microsoft

The aim here to suggesting more details to ensure you get the intended outcome. It will also offer an expanded library of ‘next prompts’.

This means if you start typing “summarise” then Copilot will display options to summarise the last 10 unread emails and chat messages or other tasks that might be related.

Prompt Rewrite

The “rewrite” feature is something that many image AI tools have had for a while. The aim is to be able to takes a person’s basic prompt rewrite it to me more thorough, “turning everyone into a prompt engineer,” according to the Microsoft.

Image (c) Microsoft

Also known as “elaborate your prompt”, Microsoft say this will be able to rewrite any prompts people create making it much easier to do more complex tasks especially when working with documents or ‘connected apps’.

Copilot Catch-up

Copilot Catch Up aims to start making Copilot more “proactive”. Here the chat interface will be able to presents people with “responsive recommendations” based on their recent activity. As an example, it will be able to notify you about upcoming meetings and suggest ways to help you prepare for that meeting, by bringing a summary of recent email and chat threads, meetings notes and documents write in the chat thread. This feature is also coming into Copilot in Outlook.

This feature brings Copilot more into the realms of good ol Clippy (ok I’m kidding here) but will enable Copilot to start proactively helping rather than waiting for its pilot to issue a command and bring the genie out of its lamp!

The aim is to further integrate Copilot into the user’s workflows. Imagine for example having a morning prompt that tells you about your day, tickets logged via Service Now, or a project that is over running (via Project or Planner) or has  completed early perhaps!

Updates to Copilot Labs

Similar to Microsoft app Prompt Buddy, Microsoft will also start to allow people to create, publish, and manage prompts in Copilot Lab.

Image (c) Microsoft

This will bring new features that that can be tailored for individual teams within businesses. This is aimed to make it a lot easier to share useful prompts for employees, Teams and departments to use.

Will these help adoption?

What do you think about the new updates, will these help remove the dark art of promoting and make Copilot easier to use and faster to help people get the desired results.?

Let me know on the comments..

Microsoft to open new AI Hub in London

Microsoft has announced plans for a new artificial intelligence (AI) hub in London, which will be focused on leading edge product development and research. This will be led Microsoft AI Lead Mustafa Suleyman (confounder of DeepMind) who Microsoft hired last month.

This annoucement comes less than a month since Microsoft unveiled a new consumer AI division.

There is an enormous pool of AI talent and expertise in the UK, and Microsoft AI plans to make a significant, long-term investment in the region. (London).

Mustafa Suleyman

This is great for the UK and for London and will help both Microsoft and the UK become an AI  and technology superpower leveraging the hub of tech talent, access to leading and world class universities and research centres with ability to attract the best talent for the next generation of development of AI.

Microsofts AI Future in the UK

This announcement builds on Microsoft’s recent commitment to invest 2.5 Billion into data centre infrastructure and improving AI skills across the UK.

Microsoft’s AI investment in the UK includes building a major new data centre in West London and installing 20,000 high-powered processors in the UK by 2026.

Microsoft’s new UK hub will be run by Jordan Hoffmann,  (another former employee from DeepMind) and will collaborate closely with OpenAI which powers Microsoft’s AI driven Copilot System framework.