39 Comments
author
Oct 15, 2022·edited Oct 15, 2022Author

Do you mean how did I merge the videos?

I've used the ffmpeg tool.

Expand full comment

“we can superimpose the beamforming data over the video” How?

Expand full comment

Hi! Could you show the code for the ReSpeaker 4 microphone array (~25 USD) with a GCC-PHAT algorithm? Please answer me,thanks!

Expand full comment

Yeah, this would be awesome! @Michael, could you please?

Expand full comment

Hi, your work is very interesting, i am also intersted in ReSpeaker and GCC Phat algorithm, can you give me some information? thank You

Expand full comment
author

Which information would you like?

Expand full comment

Having a blog post to replicate your result with the ReSpeaker 4 mic would be SUPER helpful. I already have the ReSpeaker 4 mic but found existing tutorials / code difficult to locate sound source.

Expand full comment
Jan 13, 2022·edited Jan 13, 2022

I want all the information and indications about ReSpeaker 4 to replicate your experiment, too. thanks

Expand full comment

possibly all the information and indications to replicate your experiment, thank You very much

Expand full comment

Hello, I'm also working on an Acoular project and am trying to use audio samples that have worked to debug my code. Would really appreciate the audio files if possible!

Expand full comment

Would be interesting to see what it shows when pointed at a stereo sound source. I.e. whether it will show that the sound is coming from two different speakers, or if it shows that the sound is coming from the “phantom” centre

Expand full comment

I would really like to read another blog post about the ReSpeaker 4. While looking for the shield I also found the ReSpeaker 6-Mic. Did you try your Code with the ReSpeaker 4 and a custom MicGeom file as well? How big are the differences?

Expand full comment

Are you going to fix your article to show how you superimposed the beamforming image over the video?

Expand full comment

You create an H5 format file. Then, you merge a wav and avi file. At what step did you, and how, place the overlay on the video? The article seems to have a large gap at the end.

Expand full comment

Hi Michael, this is a fantastic project. I'm wondering if you found the 4 mic array less sensitive to low frequency noise. I also wonder if this is possible with omnidirectional mics if their geometric spacing is known.

Expand full comment

Hi Mike, Do you have git repository for the ReSpeaker and GCC Phat algorithm experiment? Would be great of you put it up and possibly others to contribute if they try it and maybe refine and add to it. Just a thought as I am interested to try it and contribute. Thanks Hoo

Expand full comment

hello,Can you give me the full code for the UMA-16 DIY Acoustic Camera,thanks

Expand full comment

Can the current software capture specific frequencies? From reading your post, it seems the detection is based on the highest amplitude of the pressure wave. What if one is to introduce a tunable band-pass filter to help focus on specific frequencies only?

Expand full comment

Is there a possibility of filtering frequencies ? I want to see if specific frequencies can be isolated form same frame.

Expand full comment

m01-31-2022 Hi, there. Would you show the code for the ReSpeaker 4 microphone array (~25 USD) with a GCC-PHAT algorithm?

Expand full comment

This is very interested! I am intending to do the same thing but in 3D (hemispherical shape). These are some of the hardware so far, please allow me to share

https://photos.app.goo.gl/5zZwiGGPTdpQ1mF19

https://photos.app.goo.gl/GaWTS7Y5j7uqGkG16

Using Acoular, can we make it real time?

Thanks!

Expand full comment

When I run the record.sh script and it creates the initial audio.wav file. If I look at it using ffprobe it reports that it has captured 128 channels:

Input #0, wav, from 'audio.wav':

Duration: 00:00:19.37, bitrate: 32768 kb/s

Stream #0:0: Audio: pcm_s16le ([1][0][0][0] / 0x0001), 16000 Hz, 128 channels, s16, 32768 kb/s

[STREAM]

When I convert that to .h5 I assume it retains the same audio channels which results in an error when running the script to analyze the results.

---> 14 pm = bb.synthetic( 8000, 3 )

ValueError: 128 channels do not fit 16 mics

I'm sure this is something really simple to fix, but I'm not familiar enough to know how to alter that initial recording to only capture 16 channels of audio, 1 for each microphone so that (I assume) after convert to .h5 the script sees what it expects.

Would anyone be able to point me in the right direction? I'm running this under Ubuntu 22. Thanks in advance for any advice!

Brock

Expand full comment