Replies: 5 comments 1 reply
-
a few thoughts:
so it is a good idea but still not clear in my head how it would really be useful without some other ways of extracting meaning from it. |
Beta Was this translation helpful? Give feedback.
-
We never make a full SSM, only a rolling kernel's worth. 👎 Full SSMs are very memory intensive and pretty CPU intensive |
Beta Was this translation helpful? Give feedback.
-
This is the request that I was predicting you'd get. I'm not convinced that it isn't useful. At least I think it's another potential angle into "know your data", just the chance to see it in another way might lend some insight into what is there. Perhaps we should have a demo of doing it because people are gonna ask and then we can let them wrestle with if it is artistically valuable to them. Doing every FFT frame as a point would get too memory intensive quite quickly (a la in NoveltySlice), but the audio could be segmented differently, here's the Golcar recording broken into 1000 segments and shows the dog barks as being both more different from the other material in the recording and similar to each other (I'm riffing a bit here, but the idea is there). So, doing it in NoveltySlice, maybe not. Being able to make these matricies, and showing users how they could, might be nice. Here's the 30 minute improv set with Ben from Pittsburgh broken into 2000 segments (obviously not every FFT frame, but the time resolution here is about 0.93 seconds. BufStats gets the mean MFCC analyses over the time window, startCoeff=1). This could reveal some formal outlines or how different sections or moments relate to each other.
|
Beta Was this translation helpful? Give feedback.
-
Graphically at best, and even that. for gesture similarity, we'd need to look for diagonals that are parallel-ish (faster or slower will be steeper) but again doing that programmatically is not trivial... let's keep discussing with long-term agenda so we might inspire our late summer or someone somewhere. |
Beta Was this translation helpful? Give feedback.
-
Perhaps we can make a lower-priority help file & learn article showing how one might approach this so when users inquire about it we have a place to share some information with them. |
Beta Was this translation helpful? Give feedback.
-
After the most recent Audible Edge workshop, many participants asked if it was possible to get the self-similarity matrix out of the novelty slice object. Potential applications were suggested such as visualisation, using it as a feature for a neural network, or for trying to inspect the structure of some sound. I think the same question arose in the NOTAM workshop too.
I'm not sure if in the particular FLuCoMa implementation it is that straightforward to get it out but it might be a useful strategy for deriving a temporally sensitive feature (which are somewhat lacking ATM in the toolkit).
What are your thoughts @tremblap, @weefuzzy, and @tedmoore?
Beta Was this translation helpful? Give feedback.
All reactions