Hal 9000 voice simulator
![hal 9000 voice simulator hal 9000 voice simulator](https://3.bp.blogspot.com/_9iMIweN5ntc/SSHATZjut7I/AAAAAAAAAHI/cqMRzpBQSMk/w1200-h630-p-k-no-nu/Hal9000.jpg)
In the late 1990s, for example, Black worked with a Japanese company to synthesize Bill Clinton's voice from found data, ("I think we had him speaking Japanese," Black recalled), and when CereProc synthesized Bush's voice several years ago, they used found data from his presidential speeches.īut one big difference is that Bush didn't have a need to use the technology, whereas Ebert clearly does. "What's being done for Ebert is not completely novel," Carnegie Mellon's Black said. "Roger knows these films so well he can more or less just riff on them and a lot of this material is too spontaneous to use for synthesis."Īlthough using this found data posed a challenge, the concept wasn't new. But even then, a lot of the material was unusable, said Aylett. "A lot of the recordings contain multiple speakers, background music, audiences, different recording studios - all these things are a real problem," Aylett said.ĬereProc worked mainly with original versions of Ebert's DVD commentaries from movies such as "Casablanca," "Citizen Kane" and "Beyond the Valley of the Dolls," devoid of soundtrack or film audio.
#Hal 9000 voice simulator movie#
Although in Ebert's case, recorded material existed abundantly in the form of movie reviews and DVD commentaries, most of it wasn't ideal for voice synthesis, said Aylett. For CereProc, this meant using found - rather than targeted - data, and sifting through enough material to find all necessary speech sounds.
#Hal 9000 voice simulator software#
The software records and labels these sounds based on the original transcript and combines them into new words as needed.Įbert, however, didn't have the luxury of foresight. and in both function and content words. Also, each sound needs to appear in multiple language environments - for example, the t-sound in cat, stop, button, etc. The prompts need to contain all speech sounds necessary for the English language today - including rarer sounds like "oy," a soft "j" or sounds that appear in words of foreign origin (nasal vowels, for example). "Someone who's about to lose their voice for medical reasons can record their speech in advance," said Carnegie Mellon's Black.įor this, patients receive a transcript of about 10,000 prompts to read aloud. "Voice clones are a natural progression of our technology" of speech synthesis, Matthew Aylett, CereProc's chief technical officer, told TechNewsDaily.Īlso, voice banking has been around for more than a decade, said Carnegie Mellon's Black, and this, too, enables synthesis of a particular - rather than generic - voice.