Hi there,
last weekend I taped the concert of Kraftwerk in Bonn, Germany (08 28 2022). The concert was unseated.
I'm sorry to say, I've never experienced such an unfocused, nervous and babbling audience as at this concert. I don't know if it was due to the long Covid break, this concert was also postponed twice, but maybe people are happy just to get back among other people. But it was no fun to listen to it let alone record it.
About three quarters of an hour before the show I looked for a good standing place in the area in front of the stage. Shortly before the concert I then noticed at some point behind me a group of (sorry) "housewives" who had obviously strayed to this concert of Kraftwerk. Then the crowd concentrated in the area in front of the stage, where I was standing. At some point, however, it was so full that I no longer had the opportunity to place myself somewhere else. This group of women behind me actually talked loudly for the whole two hours, as if they were sitting in a cafeteria with background music. Especially in the quiet passages this was unbearable - in the loud passages was then of course shouted.
All the babbling can of course be heard on the recording, and in such a way that it is really disturbing.
This recording has now led me to deal with the attenuation or elimination of audience noise. I haven't even tried that before because I thought it would be quite difficult and time consuming or almost impossible.
I own the program iZotope RX 9 and experimented a bit with it. My idea behind it was: If there is the possibility to isolate only the vocals (or human sounds), then it should be possible to get a track without vocals in the opposite sense. However, with the tools "Dialogue Isolate" and "Music Rebalance" this worked only moderately, even if I only select the relevant areas in the WAV file. The tools does indeed find human sounds, but only very litte.
Then I came across the website of
www.lalal.ai. These relatively new possibilities to split tracks into the different instruments seem to be quite popular with DJs at the moment.
I then uploaded a few samples there with a blatant example and I have to say that I was quite flabbergasted at how good the algorithm used there is. The babble was filtered very well, except for a few pieces of music.
Sure, you can't use everything that this algorithm then renders from the files, since of course it also includes the vocals, and there will still be a lot of editing work to be done, but unfortunately the prices for a completely rendered concert are not what I like either imagine.
Has anyone here had similar or good experiences with eliminating audience noise, especially with these new methods of artificial intelligence?
Cheers
Michael