-
jermain
adc~ has no signal when monitored with [env~] or [pvu~] --> [vu].
The internal microphone works. Audio drivers on Pd are set to take Built in microphone 2 channels. Audio and Midi drivers (on a Mac) is set to use internal microphone. I've used system preferences to test/playback mic input.
Other sounds work in Pd. DSP is on/computing. I tried listening for anything despite the visual showing that it wasn't working -- with headphones. I'm at a loss. Any help here is appreciated.
-
jermain
This is an old post but for anyone who may have ended up reading I found a useful workaround with Soundflower. It's a free plugin for routing audio between programs, available through Rogue Amoeba and originally through Cycling '74
-
jermain
I noticed this and signed up a while back. Do these still meet, anybody know? Sounds like a cool idea.
-
jermain
http://www.pd-tutorial.com/english/ch03s09.html
The chapter invites me to build an expander. I'm not sure I get how the math concepts in this chapter can be applied to make one. Is it as simple as switching some of the operations (plus for minus) in the compressor patch?
-
jermain
Calling all engineers!
Is there anyone who can get me started on my way to physical modeling? I would just like someone to provide in laymen's terms the basic conceptual difference between creating a software instrument by modeling it vsd FM synthesis.
Or is FM synthesis + studying spectrograms of instrument samples and using bp filters, noise bursts, etc. to match them the extent of it? (I've tried it so far to little success).
A generic answer is desirable, but just so you know I am going to attempt to use PD to make acoustic piano sounds as a first project. Maybe someday I can have a whole software orchestra do Mahler, why not?
P.S. If anyone has not looked into Yamaha's AN, here is what inspired the project:
-
jermain
Hi all,
I was working on some speech synthesis in PD by emulating spectral analysis of speech. This was doable. But I got curious about emulating other types of speech, like screaming and singing.
Any advice on how to accomplish this? I used fft~ to analyze my own screams, and did some searches for analysis already done online. There were a couple papers available.
I made some experiments to simulate screaming already. The fft showed that unlike a normal speaking voice which shows only a few spectral peaks and is convincingly mimicked with a few bp filters, spectral peaks for screaming are active throughout the audible spectrum.
First I tried setting up a regular formant synth with higher fundamental tones, setting the gain higher on the partials and adding in some noise. This sounded too much like whispering.
Second I created parallel bp filters for the noise to try and make it less airy and more direct/forceful sounding.
Third I tried to skip the noise entirely and simply add more [phasor~] objects where I saw partials. There had to be so many that they tended to cancel out or sound way too distorted to be like a speaking or screaming voice.
I'm open to any kind of response, theoretical or practical. Just can't wrap my head around the problem yet.
-
jermain
I am interested in extracting voices (instruments or vocal) from a song - an idea I got from an old MIT lecture on DSP I saw on YouTube (they're pretty cool and easy to look up so I recommend them). I didn't see anything on the forums. I am a noob with DSP theory. Right now I am working through Andy Farnell's and Miller Puckette's books.
I will be grateful if someone could point me to methods or resources that address this kind of processing. I know there is no quick fix and I will probably have to study on this problem for a while. Just need some help getting started!
If this idea is possible to implement the most helpful response would be if someone could define the problem for me. The article attached is the first result on my Google search and I thought even as a layman I could mine it for a possible solution.
Finally I'm wondering if this processing technique can be done in real time.
Thanks for looking at my post.
-
jermain
Thanks nau, I searched the forums first but I hadn't encountered that one.
Maelstorm - great link, thank you. Comprehensive and just what I was looking for . . . DSP heaven!
-
jermain
You're exactly right. Thanks for the concise explanation, Maelstorm. It's actually not a first project, just the first one where I am experimenting with a precise idea of the sound I want to end up with.
It sounds very complex, but would you say more? What kind of homework is involved in this problem, assuming a knowledge of physics and calculus? (It doesn't matter to me if it takes some considerable research.)
-
jermain
acreil, I did a search of 'false vocal chords'. It definitely seems right that they play a role. If they are thicker than true vocal chords perhaps they should have their own timbre, maybe warmer sounding. Perhaps if they are meant to interact then I should add them in series rather than parallel, and add feedback? Also, thanks for the throat singing example!
Adding two phasors worked pretty well, it actually produced its own distortion. I haven't worked on this in a while, but today I played around with some osc~ based amplitude modulation after my formant filters and it added some realism.
Flipp, I found out there is actually an external for speech synthesis.
http://puredata.hurleur.com/sujet-2092-speech-synthesisI couldn't build it on my Mac, though it's apparently possible. But I wasn't actually thinking of going in this direction simply because I don't have enough programming experience. I just wanted to create some vowel sounds, or partial words.
-
jermain
Actually someone else on the forums discovered that adding noise makes whispering sounds. I don't have it in front of me right now so I can't credit the creator, but I found an incredibly nice abstraction of a formant synth that was my jumping off point for this project. It's the one that includes Turkish vowels. Comes out pretty convincing - sounds like the whisper voice on Mac.
Many thanks for this article, I had not seen it.
Here is an intriguing article I referenced that analyzes death metal growling, if you or anyone is interested:
http://courses.physics.illinois.edu/phys406/Student_Projects/Spring05/Chuck_Stelzner/Chuck_Stelzner_P498POM_Final_Report.pdf -
jermain
I wonder if you've seen this external library.
http://williambrent.conflations.com/pages/research.html
I was thinking of timbre ID. It has a nice speech recognition object. I used it (on Mac OSX Lion - no issues) on a project where I sent speech simultaneously to [fiddle~] and the speech recognition and used them simultaneously to control the parameters of a synth. Which sounds fairly close to the type of project you're doing.
These are the ready to hand solutions I can think of. But if you want to build it yourself I believe you could begin by looking at the vocoder and timbre stamp patches in the PD help browser (I06. and I07.)
Also, if you need to make one audio signal sound like another (not sure if this is what you mean or importantly if it should be real-time) then [convolve] from the same library I mentioned or the timbre based granular synth could work well.