Creating a Modular Synth - Native Implementation, or Middleware

So – I’m working on building a modular synth in Unity, and so far I’ve written a multi-voice oscillator, and basic controller logic to play keys on your keyboard.

Do you think there’s much benefit to writing it all natively? I’m interested in learning some middleware regardless, but I can’t seem to find much information as to whether a modular approach to synthesis would be feasible using middleware.

Looking for input and insight into a project like this.

Also - currently I’m playing with OnAudioFilterRead, and it seems generally a little latent. Is there an approach to directly manipulating audio samples that might make a bit more sense?

Open to any and all discussion about this :smile:

We’ve had a few conversations on that matter, this one is a relevant recent one .

TLDR: Unity by itself lacks the performance required for dedicated audio applications, in which case native solutions are recommended. For projects not targeting a professional audio audience, it is still a relevant choice though!