I would like to get a live monitor of the output of an ffmpeg process (it's
actually a capture and compress process).
I believe I could achieve this by adding a streaming output, and then
separately starting an ffplay process to display that (I haven't tried this
yet, but have got most of the elements working separately at least).
However, this network layer (even as local loopback) seems like a small,
but undesirable, overhead. Can anyone tell me: a) is my concern entirely
unfounded? and b) is it possible to do this more efficiently, presumably in
the main ffmpeg process? and c) is there some potential benefit to the
local-network-between-two-separate-processes approach that I've not thought