So – I’m working on building a modular synth in Unity, and so far I’ve written a multi-voice oscillator, and basic controller logic to play keys on your keyboard.
Do you think there’s much benefit to writing it all natively? I’m interested in learning some middleware regardless, but I can’t seem to find much information as to whether a modular approach to synthesis would be feasible using middleware.
Looking for input and insight into a project like this.
Also - currently I’m playing with OnAudioFilterRead, and it seems generally a little latent. Is there an approach to directly manipulating audio samples that might make a bit more sense?
Open to any and all discussion about this