Stanford scientists have developed a system that can automatically synthesise realistic sounds for
In addition to enlivening movies and virtual reality worlds, the system could also help engineering
companies prototype how products would sound before being physically produced.
It could also encourage designs that are quieter and less irritating, the researchers at Stanford University
in the US said. “The first water sounds we generated with the system were among the best ones we had
simulated — and water is a huge challenge in computer-generated sound,” said Doug James, professor
at Stanford University.
Informed by geometry and physical motion, the system figures out the vibrations of each object and
how, like a loudspeaker, those vibrations excite sound waves.
It computes the pressure waves cast off by rapidly moving and vibrating surfaces but does not replicate
Although it does not recreate the echoes in a grand cathedral, it can resolve detailed sounds from
scenarios like a crashing cymbal, an upside-down bowl spinning to a stop, a glass filling up with water or
a virtual character talking into a megaphone.
Most sounds associated with animations rely on pre-recorded clips, which require vast manual effort to
synchronise with the action on-screen.
These clips are also restricted to noises that exist — they can’t predict anything new.
“Ours is essentially just a render button with minimal pre-processing that treats all objects together in
one acoustic wave simulation,” said Ante Qu, a graduate student Stanford University. The simulated
sound that results from this method is highly detailed.
It takes into account the sound waves produced by each object in an animation but also predicts how
those waves bend, bounce or deaden based on their interactions with other objects and sound waves in
the scene. Source: PTI