I have a project that uses a proprietary SDK for decoding raw video. I output the decoded data as pure RGBA in a way FFMpeg can read through a pipe to re-encode the video to a standard codec. FFMpeg can't include the Non-Free SDK in their source, and it would be wildly impracticable to store the pure RGBA in a file. So pipes are the only way to do it, there are valid reasons to use high throughput pipes.
ploxiln|1 year ago
CyberDildonics|1 year ago
Lets not get carried away. You can use ffmpeg as a library and encode buffers in a few dozen lines of C++.
quietbritishjim|1 year ago
whiterknight|1 year ago
Almondsetat|1 year ago
whartung|1 year ago
It's clumsier, to be sure, but if performance is your goal, the socket should be faster.
AkBKukU|1 year ago
[1] https://ffmpeg.org/ffmpeg-protocols.html#unix
ptx|1 year ago
jcelerier|1 year ago
Sesse__|1 year ago
jraph|1 year ago
But then the solutions are not comparable anymore, are they? Would a lossless codec instead have improved speed?