DE · Topics · Design · NVH

Is Your Car a Good Listener?

Breaking down the role of acoustics analysis in electric and autonomous vehicle development.

Breaking down the role of acoustics analysis in electric and autonomous vehicle development.

View or stream online

Audio Podcast
Is Your Car a Good Listener?
Duration
9:11 hrs/min/sec
View or Stream

Subscribe via iTunes or Google Podcasts and don't miss a beat.

Subscribe today


Subscribe to our DE247 Podcast Channel and get engineering news and content delivered in real time via Apple iTunes, Google Podcasts and more.


Human drivers are trained to rely on “visibles” and “audibles” to make split-second decisions. An emergency vehicle’s siren, unusual mechanical noises from the engine and the loud bells from railroad crossings are just a few of the auditory alerts that might prompt a sudden change in driving strategy. So, when the human driver is removed from much of the decision-making process in autonomous vehicles, who is listening? 

Your car is, as it turns out. There’s a specific discipline that studies the noise, vibration and harshness (NVH) of the vehicle. In classic automotive design, the noise-refinement goal is primarily to reduce the level of undesirable noises in the car’s interior. If your engine noise is too loud, you won’t be able to enjoy the latest Taylor Swift single streaming from your smartphone, for example. But with the emergence of self-driving cars, the discipline takes on new burdens: one is teaching the car to listen.

Listen to the recorded podcast with the speakers quoted below:

Simulation of sound propagation inside a car’s cabin. Image courtesy of COMSOL.

A Multiphysics Approach

Mads Jensen, technical product manager for Acoustics at COMSOL, is someone who listens to your future car’s hums and whispers. He points out the reversal of acoustic engineering—the shift from noise cancellation to noise amplification—with the emergence of electric vehicles.

“The autonomous car itself has to react to the environment. ... That means the microphones in the car have to be placed properly to hear the environment better.” 

— Siva Senthooran, director of technical sales at SIMULIA Centers of Excellence.​​​​​​

“In different countries and regions, the car is required to make a certain level of sound so the pedestrians can hear it approaching. So, with electric cars [which do not have a noisy combustion engine], you now have to actually emit sound with loudspeakers to meet that requirement,” explains Jensen.

Modeling the outbound acoustics of the car, such as sound emitted from the car’s speakers, can be done with the Acoustic Module in the COMSOL Multiphysics software suite. “Acoustics can be coupled with other physical effects, including structural mechanics, piezoelectricity [accumulated electric charge] and fluid flow,” according to COMSOL’s product details page. 

“A lot of the car makers are modeling the acoustics inside the car as well. It’s a relatively small acoustic space, but a complex environment. You have people and objects, so you have to use multiple speakers to create sound zones inside the space,” Jensen says.

Imagine a fully autonomous van in which the driver has the option to turn around for chitchat with the passengers sitting in the back. The optimal speaker placements for such a vehicle are significantly different from a typical family van where everyone is facing the front for the most part. 

“To figure out the optimal configuration of speakers and sound zones for such a space, you really need to rely on simulation,” adds Jensen.

By itself, sound propagation is not multiphysics; however, the mixture of electromagnetics, structural vibration and sound fields to account for the behavior of the speakers makes the simulation multiphysics—a specialty of COMSOL’s simulation technology, Jensen points out.

Though electric cars do not have a combustion engine, they do have an electric motor, which generates certain noise and vibration. “These sounds can propagate through the vehicle’s body to create unwanted sound inside the cabin,” Jensen says. “When you have sound fields interacting with the car’s structure, it’s a multiphysics problem.”

Raytracing for Sound Simulation

Engineers traditionally use finite element analysis (FEA) methods to model the wave fields inside the vehicle and solve simulation in the frequency domain or time domain, Jensen explains. 

“With this approach, you need a lot of discrete points or mesh elements to resolve the wavelengths, so it’s computationally expensive … With the latest development in COMSOL software, you can now run the full wave approach to a much higher frequency. Both computational methods and computational power are improving,” says Jensen. 

An improvement in the software is the use of time-explicit discontinuous Galerkin method. “This allows you to model wave propagation in the time domain,” says Jensen. “This approach is extremely suitable for parallel computing.” 

The latest strategy is to employ a hybrid approach—a mix of full wave simulation and raytracing. “With raytracing, you can think of soundwaves as particles traveling like rays. When it hits a surface, it bounces back, but some energy is also absorbed by the surface, depending on the surface’s porous or smooth texture,” says Jensen. “That’s an inexpensive way to model acoustics inside a confined space.”

An important step in the acoustic simulation is the treatment of the CAD geometry, which defines the space in which soundwaves propagate. 

“The better you can prepare your CAD geometry for meshing and simulation, the more accurate your simulation is,” says Jensen.

Too Quiet for Safety

Ales Alajbegovic, vice president of worldwide industry process success & services at Dassault Systèmes’ SIMULIA, recalled a cab ride to the Munich International Airport. The cab driver was complaining about the loud wind noise, but the wind was not particularly strong that day. 

“When you remove the noise of the combustion engine, the noise inside the cabin is reduced significantly. But as a result, you start to notice other noises, like wind noises, and much more,” he observes. “In many cases, they may be even more bothersome than a car with a combustion engine.”

Wind noise simulation showing turbulence and the resulting acoustic waves on glass panels. Image courtesy of Dassault Systèmes.

Wind noise comes from aerodynamics, or the friction between the airflow around the car and the car’s chassis itself. Therefore, “the shape of the car is crucial to what you hear in the cabin,” says Alajbegovic. One way to minimize it is to reduce drag by modifying the shape of the chassis—the geometry of the car. 

“The autonomous car itself has to react to the environment. So, in addition to seeing [using real-time video feed and computer vision], it also has to listen to what’s happening around it. That means the microphones in the car have to be placed properly to hear the environment better,” says Siva Senthooran, director of technical sales at SIMULIA Centers of Excellence.

Dassault Systèmes offers several tools for aerodynamic simulation and geometry modeling in its 3DEXPERIENCE product line. PowerFLOW software, based on transient Lattice Boltzmann-based physics, includes features for thermal, aerodynamic and aeroacoustic simulation. The company also offers Wave6 software for vibro-acoustic and aero-vibro-acoustic simulation; and Abacus, a structural analysis suite. 

Can You Hear Your Car?

In May 2018, Ansys acquired OPTIS, an optical simulation software developer. Subsequently, Ansys also acquired GENESIS, an acoustic simulation software developer. The new acquisition enables Ansys software to generate not only visual representation of sound fields but also the audio—in theory, what you would hear inside the vehicle.

“The electric vehicles are quiet—maybe too quiet. So secondary noises like hatchback noise and duct effects become clearly audible. Also, the electric motor has a fast rotation, which creates a whining sound,” says Patrick Boussard, senior manager, product management at Ansys. Boussard is also the founder of GENESIS before its acquisition. 

“Because the car is too quiet, any squeak or creak from the panels or misalignment in the windows gets noticed. There’s nothing to drown out these sounds,” says Siddharth Shah, principal product manager, Ansys.

Acoustic analysis gets more complex in combining autonomous driving and the quieter electric powertrains. With the need to focus less, humans in the autonomous vehicles are more likely to listen to music, watch a movie, engage in conversation or even take a nap. That means the vehicle needs to create a variety of sounds to alert the passengers to lane changes, emergency stops and the need for someone to assume manual control.

Using Ansys VRXPERIENCE Sound, users can virtually experience the cabin of a car still in development, visually and auditorily. 

“It adds signal processing, so you can hear all the sounds from different audio sources in a surround environment,” explains Boussard. 

“What you can do with VRXPERIENCE is to assess the sound,” says Shah. “With it, you can tell whether you increase the sound by a certain decibel if you make changes to the mirror, the vent or the profile of the car, for example.”

In autonomous cars, the in-cabin audio experience takes on new importance as drivers and passengers may demand a better environment for entertainment (movies, music and conversation, for example). But the car must now also learn to listen and react to the sound signals the drivers may not be actively listening to.

View or Stream

More Ansys Coverage

More COMSOL Coverage

COMSOL Company Profile

More Dassault Systemes Coverage

Unified Modeling And Simulation (MODSIM) For Sustainable Product Development
Companies in all sectors must address the need for greater sustainability to meet customer demands. The development of new products necessitates rigorous testing and evaluation, while efforts must be made to decrease emissions and operate in a more economical fashion.
What’s New in 3D Motion Creator R2024x
3D Motion Creator provides simple functionality for kinematic and dynamic motion analysis of assemblies.
Industry Leaders Shift Strategies to Harvest AI Spring
AI-powered simulation and natural language input poised to become the norm.
Dassault Systèmes Releases SOLIDWORKS 2024
Company says new release enables users to create experiences smarter, faster, together.
Twin Tech and the Smart City: A Natural Pairing?
The question is whether the digital twin can take the smart city into the mainstream.
US DoE Puts $2.85 Million Toward 3D Printing Modular Wind Blades
Funding was awarded to Purdue University and its industry partners, including Thermwood, TPI Composites Inc. and Dassault Systèmes.
Dassault Systemes Company Profile

Share This Article

Subscribe to our FREE magazine, FREE email newsletters or both!

Join over 90,000 engineering professionals who get fresh engineering news as soon as it is published.


About the Author

Kenneth Wong's avatar
Kenneth Wong

Kenneth Wong is Digital Engineering’s resident blogger and senior editor. Email him at kennethwong@digitaleng.news or share your thoughts on this article at digitaleng.news/facebook.

      Follow DE
#24396