Hi, First time poster on this forum but I’ve used Audacity every now and again over many years.
This post is primarily a response to Tantacruls video (https://youtu.be/RMWNvwLiXIQ) to provide feedback on where I’ve experienced pain with Audacity before and where I think it can be improved.
Feedback:
For simpler audio editing tasks, I’ve found myself using https://www.ocenaudio.com/ instead due its more responsive UI and cleaner design. But it lacks the more advanced features that Audacity has.
The most common pain I experienced was the UX loop of “apply effect, listen to it, slightly wrong parameters, undo, repeat”. And the “preview” often wasn’t helpful.
I would love it if we had like the stackable effects shown In Jessica Williamson’s mockup (https://youtu.be/RMWNvwLiXIQ?t=791) which would allow tweaking of parameters while I’m continuously listening to a second of audio. This seems to tie in with this popular feature request https://wiki.audacityteam.org/wiki/Feature_Requests#realtime
I would advocate for a feature that is more general which I found very useful in other applications which deal with composing slow and real-time effects together: Procedural editing.
The key difference is that it allows users to instead of changing the content, to compose nodes that operate. I’ll give two examples that I think demonstrate this well:
Houdini (commercial) https://youtu.be/Tsv8UGqDibc?t=641
Blender (open source) https://youtu.be/Uvc2P911rZk
Some reasons why I really like this approach to UX:
- It is non-destructive: the original samples are kept and rendered audio can be regenerated.
- Users can easily do a with/without effect comparison even after other actions have happened (which would occupy the Ctrl+z history)
- Certain things that have their own special UI currently like “Solo” would be a natural thing like viewing the track node before it gets added to other tracks
- Project files can be small as the generated data does not need to be saved.
- Nodes are general and could span multiple audio tracks such for instance, audio between two tracks could be multiplied together.
- Allows organization of complex chains into blocks that can be reasoned about and shared to other users (like functions in programming)
- Can be real-time nodes by default and control can be given to users when they want to “cache” the output of a node to avoid long re-computations of effects.
- I think it maps cleanly to a plugin API
Note that I think this scales quite nicely with the complexity of what the user is trying to do.
If it’s simple → graph is simple and the graph view could be hidden by default and actions on the tracks (like currently) just generate nodes in the background.
If it’s complex → graph allows to manage the complexity.
I’m a programmer by day and I’d be willing to put in time to make it happen if you decide to go down the procedural editing route.
I’ve never really used the per-track audio volume and panning. Maybe these could be moved to being a real-time effect as part of the above?
I’ve never really needed the description of “stereo, 44100Hzm 32-bit float”, I think that can be put in an info dialog.
The Spectogram feature is useful and I thought it was difficult to find as I was searching in the “View” menu rather than knowing that it was a per-track thing.
When I select a rectangular region of a spectogram (not the full height) and press delete, I’d expect the selected frequencies to be dropped to zero in the selected time frame, rather than the time frame being completely cut. I ended up removing things with effects but that was a bit painful. Also it would be nice if a frequency selection is made and I hit play, that I hear only those frequencies. This would’ve been useful when I was trying to isolate information from an audio file that came from from an ARG.
I like dark modes for all my applications.
Logo: Keep the icon looking the same/similar; I don’t mind if the text style changes.
I’d be open to be interviewed as a user if you’d like further details.
Hope this feedback is constructive. Regards,
:Dom