ReadSpeaker and Audiokinetic Introduce speechEngine for Wwise: Runtime On-Device Text to Speech

Wwise Tips & Tools

Audiokinetic and ReadSpeaker are excited to bring a deep integration of speechEngine into the Wwise pipeline, allowing developers to harness the power and flexibility of on-device text to speech in their game. speechEngine for Wwise is a cross-platform solution and seamlessly integrates into the game development workflow. 

speechEngine's runtime capabilities give developers full control, enabling real-time conversion of in-game text to speech. Whether they’re looking to enhance accessibility with UI narration and audio descriptions, add speech to multiplayer interactions, or build narrative experiences, speechEngine will enable developers to add dynamic, natural-sounding speech to their game.

 

On-device Processing

The advantage of speechEngine lies in its on-device capabilities. By using a CPU-based computing framework we are able to synthesize speech locally on the device. This means that there is no network connection required in any step of the process. Using optimized algorithms for inference, along with lightweight models, speechEngine produces speech on a single thread of the CPU with fast response times. At the same time, we are able to keep the footprint of our technology small, with each voice engine requiring ~10-15 MB of memory. Additionally, we have made this technology available across platforms, allowing cross-platform deployment.

Author, Export and Dynamically Adjust

The speechEngine integration for Wwise adds voice control parameters directly to the Authoring interface, enabling fine tuning of the voice profile along with the vast range of effects provided by Wwise to further customize the sound of the voice. When exporting to a SoundBank, you control the voice engine at runtime. By sending text input to the plug-in through the Wwise SDK and playing the voice, speechEngine immediately reads the text aloud. Furthermore, you can access controls for pitch, speed and more utilizing RTPCs. For more fine-grained control, Speech Synthesis Markup Language (SSML) can be used as text input for further customization.

speechengine-voice-Wwise-plugin speechEngine-voice-Wwise-choose

Speech Synthesis Markup Language (SSML)

SSML allows you to further customize how text will be read. Using the simple markup language syntax of SSML you can achieve a multitude of effects, such as putting emphasis on a certain part of the speech, like in these examples:

Example 1: "Listen carefully!"

<emphasis level=”strong”>Listen carefully!</emphasis>

Kayla
Molly
Sophie
 
 

 

 

 

 

Example 2: Inserting breaks

We are in dangerous territory <break time=”200ms”> proceed carefully from here on

Kayla
Molly
Sophie
 
 

 

 

 

 

Example 3: Change how the text is interpreted

<say-as interpret-as="characters">wasd</say-as>

Kayla
Molly
Sophie
 
 

 

 

 

 

Example 4: Spelling words out phoneme by phoneme:

Hello <phoneme alphabet="ipa" ph="wɜːld">world</phoneme>

Kayla
Molly
Sophie
 
 

 

 

 

 

For details on the full range of SSML possibilities, see the official W3C specification: https://www.w3.org/TR/speech-synthesis/#S3.2.3 

Voice Library

ReadSpeaker offers a broad range of voice options, with a library of over 115 voices in more than 40 languages. The initial speechEngine for Wwise plug-in release ships with 12 unique-sounding voices in 10 different languages. We plan to continue to expand the selection of languages and voices in future releases.

What Will You Create With speechEngine?

We invite game developers to explore the power of speechEngine for Wwise. Connect with us on Discord or drop us an email (gaming@readspeaker.com). 

ReadSpeaker is committed to delivering ethical AI-powered voice solutions. Learn more about their approach to ethical AI voice in gaming here.

ReadSpeaker

ReadSpeaker

ReadSpeaker is a global voice specialist providing dozens of languages and lifelike voices.

ReadSpeaker's speechEngine for Wwise is a cross-platform, on-device text-to-speech solution for creating natural-sounding speech from any text input.

Comments

Leave a Reply

Your email address will not be published.

More articles

Simplifying WAAPI

If you’ve never used the Wwise Authoring API (WAAPI) before, I’m hoping I can convince you to give...

29.5.2018 - By Adam T. Croft

Event-Based Packaging Process Overview

What is Event-Based Packaging? Not long ago, the UE4 Integration of Wwise 2019.2 launched a new...

21.1.2021 - By Fan Runpeng

Wwise Beginners: 10 Questions Answered

Learning a new software can be a challenging process, yet exciting and rewarding at the same time....

6.3.2024 - By Mads Maretty Sønderup

Realistic Haptics Effects Achieved With Wwise

Introduction My name is Natsuo Koda from Miraisens. I am the CEO of Miraisens, engaged in R&D...

3.4.2024 - By Natsuo Koda

Wwise Release Cycle News | Simultaneous Patch Releases & Development Support for Unreal Engine Preview

This post aims to share some changes we made to our development process in the past few months....

14.5.2024 - By Guillaume Renaud

Music Render: Offline Rendering Tool

Implement music more efficiently When implementing game music using Wwise, you need to play back the...

17.7.2025 - By Naoto Tanaka

More articles

Simplifying WAAPI

If you’ve never used the Wwise Authoring API (WAAPI) before, I’m hoping I can convince you to give...

Event-Based Packaging Process Overview

What is Event-Based Packaging? Not long ago, the UE4 Integration of Wwise 2019.2 launched a new...

Wwise Beginners: 10 Questions Answered

Learning a new software can be a challenging process, yet exciting and rewarding at the same time....