Video conferencing from the Titanic wreck?

Thanks to this researcher, there's literally no place to hide from Zoom calls.
Written by Greg Nichols, Contributing Writer
OceanGate Expeditions

A researcher has brought video conferencing technology to one of the most remote places on earth: The wreck of the HMS Titanic, which is resting on the seabed 13,000 feet below the surface.

"It is as if we can now carry out video conferences from the abyss," says Alex Waibel, a researcher at Carnegie Mellon University and Karlsruhe Institute of Technology.

Scared yet?

Waibel is an expert in text to speech technology. Currently, the only way for researchers exploring the Titanic wreck or other deep sea targets in submersibles to communicate with the surface is via text messages sent by sonar. Radio signals don't work well underwater, presenting a communications quandary that scientists have been finding workarounds for since WWII.

During a recent OceanGate Expeditions voyage, Waibel narrated his dive and used speech recognition technology to convert what he was saying to transmittable messages. On the surface, the technology Waibel and his team pioneered then resynthesized the crude text messages to video using AI. The result was a near real-time video that used Waibel's voice over a video that looked like his lips moving in sync with the words. These efforts are aimed at aiding natural communication in extreme environments but could have potential in consumer markets as well. Waibel is a Zoom research fellow and advises the company's AI research and language technology development. 

"By interpreting and recreating natural voice communication, we are trying to reduce the workload of  scientists and pilots in such missions in a natural way, despite the challenges imposed by salt water, operational stress, conversational dialogue and poor acoustic condition," Waibel told CMU's Aaron Aupperlee.

We've written about the tremendous advances and market growth of speech recognition, which is entering an accelerated phase of development and adoption across a number of key sectors. Waibel's work builds on that trend with a delivery mechanism that uses low bandwidth broadcasts (in this case by sonar) to effectively deliver full, albeit synthesized, video to the end user.

The technology uses a synthesized voice that sounds like the speaker, building on advances in AI-powered text to speech technology. One other potential application of the technology is rapid translation from one language to another, where an end user sees a video in a comprehensible language that the speaker doesn't actually know. 

Editorial standards