Especially since this states H.264 compliance, I'd be willing to say this can't possibly live up to it's claims.
Video encoders are notorious for their bloated claims and faulty comparisons. After a quick perusal of their site I see they compare themselves only to web-m, and use generalized %'s to claim differences between products and processes.
They say they've optimized for human perceptual quality. Perhaps they have. But did they really do a better job than the open source H.264 alternative (x264) used by everyone from YouTube and Hulu, to Facebook? One thats been in development for _years_ now and beats just about everything thrown at it in a fair comparison?
I think not. I'd love to be wrong though.
So their encoder is based on x264 (without credit). The settings are slightly different between the clips, so maybe they have some magic thing to detect which settings to use. In any case, it'd be interesting to reencode the original clip using plain x264, once using the same options they used and once using some sane x264 preset.
>In any case, it'd be interesting to reencode the original clip using plain x264, once using the same options they used and once using some sane x264 preset.
I'm up for that. It'll take a while to download all the source videos, but I'll report back with results and videos when I'm done.
I already did a test encode with x264 --preset veryslow --tune film --level 4.0 --crf 21 on the Bourne Identity clip, and the result is within the same realm than the Beamr video with ~58% the filesize. Which is not surprising at all, considering Beamr's settings.
EDIT: Testing with nearly identical settings (x264 --bframes 0 -- subme 8 --no-psy --no-mbtree) using the latest x264 gives me about 40% the bitrate of the Beamr example for the Bourne Identity clip, which unsurprisingly looks notably worse. So maybe Beamr's "magical" encoding technology is all about fiddling with x264's CRF to make it double the bitrate at a given value compared to vanilla x264?
Just to note, no credit is necessarily required. x264 is now dual-licensed under GPL2 + commercial license. If they bought a commercial license, there's nothing illegal or wrong in what they do.
Incremental improvements in H.264 are still possible but not a halving of bitrate in the general case. The required rate depends very much on the content of the video and the size at which it will be viewed.
Any encoder manufacturer can find some great looking and easy to encode images that can be massively reduced in data rate without much visible impact. However in tougher scenarios they will be unlikely to beat the existing market leaders by more than 10% or so under similar constraints (latentcy/CPU/multi-pass possible) under the review of experts. Tough scenarios will be when there is lots of noise and non-uniform movement. Pop concerts can be very challenging with crowd movements, flashing lights and low general light levels leading to noise.
The more detailed claim is 5-6Mbps streams can be reduced to 3-4, so it's more like a 25% improvement methinks.
Their first claimed success was with JPEGmini — this is a lot more credible since JPEG is a pretty obsolete format that simply continues because it's so well supported. JPEG2000 delivers the purported benefits of JPEGmini for real -- it's just poorly supported.
I suspect all they're really doing is fine-tuning compression settings based on differencing input and output using a decent perceptual algorithm.
I uploaded a photo straight out of a Nikon V1 onto JPEGmini and got a 2.2x reduction. I could do better than that by dialing down quality until I saw a difference. JPEG fine is way overkill for most purposes, and that's their market.
As others have noted, Beamr's "magical" technology simply seems to be using x264, the state-of-the-art in H.264 encoders, which also happens to be free and open source software. Looking at their site, beamrvideo.com, the bitrates they list for the originals are what you'll typically find on Blu-ray discs, not on online streams. Anyone who knows a thing or two about video encoding and x264 will know that you can re-encode these to a much smaller size while retaining most of the visual quality. As such, claiming that they can reduce the bitrate to half with no notable loss in quality is like claiming that you can reduce the bit depth and sampling rate of a 24-bit 192 kHz audio file to 16-bit 48 kHz with no notable loss in quality - it's certainly true, but you hardly needed the 24-bit 192 kHz quality to begin with.
They don't seem to have done anything new or substantial by themselves, and as such, this thing is basically nothing but pure marketing bullshit. If you want to do high quality H.264 encodes, just use x264 directly.
I grabbed 4th pair of their demo image [0], the high quality image is about 4.8M, the JpegMini'd version - 1.3M.
Then I took the high quality image and started bumping up standard JPEG compression level until I reduced the file size to 1.3M. That mapped to 17/100 setting, with 1 being highest quality, 100 - highest compression).
Then I made an arithmetic per-pixel diff between the original high quality image and two 1.3M versions. Cropped them to 2000 x 2000 and saved as PNGs (cropping is to fit the imgur file size limits). Here they are -
Haha, I only kind of skimmed your post then downloaded that file. I opened the images and was like "I don't know how high-quality this is. The images look exactly the same to me." And then I saw the file sizes and had the ah-ha moment.
Maybe I'm not the most knowledgeable guy here, but I really don't see what you're getting at. The per-pixel diffs look nearly the same. I downloaded the "hipnshoot.zip" and yes the photos look identical. But then I fired up gimp and re-compressed the high quality photo to reduce the size even more than the "jpegmini" version, and I can't tell the difference. Here are the file sizes:
-rw-r--r-- 1 tim users 4930482 Dec 1 21:22 hipnshoot.jpg
-rw-r--r-- 1 tim users 1310430 Feb 26 08:27 hipnshoot_gimp.jpg
-rw-r--r-- 1 tim users 1352882 Dec 3 09:16 hipnshoot_mini.jpg
I wrote a JPEG encoder once, and if I remember correctly you can change the quantization tables and the Huffman tables. My encoder just used the default tables in the standard, but it wouldn't surprise me to optimize these using a psychovisual fitness function.
>Elements of the recordings that are easily perceived are represented with exacting precision, while other parts that are not very audible can be represented less accurately. Meanwhile, inaudible information can be discarded altogether.
AFAIK JPEGmini has much better psychovisual optimization than libjpeg, so I think their claim that they can compress JPEG better within limits of the format is fair.
The way you encode lossy formats makes a lot of difference. For example the same x264 encoder gives dramatically worse results when aiming for optimal PSNR rather than using its psychovisual optimizer:
Let them put up a long, complex video -- an action movie, for example -- using H.264, and their own method, for side-by-side comparison, both with regard to quality and file size. Let them offer something better than hand-waving press releases.
I say this because claims like this tend to be very source-dependent. Change the source, change the claim.
They appear to be using modern x264. This does outperform most (all?) commercial H.264 encoders to the degree they claim, so their claims can have merit.
However, this also means that if you want better encoding, you can just grab x264 and do it yourself.
So, in the case of jpegs, their compression is not lossless, they just claim the losses are not that perceptually visible.
Apart from the strong PR claims, this is actually not that bad, no? If you have two lossy compressions, but one is perceptually better, you will rather use the perceptually better one, no?
Just using x264 with veryslow and/or two pass setting usually cuts file sizes by half compared to e.g. an IP or phone camera output. They could have some magic there, but you can get similar results at home without having to pay anyone. You just have to read a little about how to run x264.
So, since it's seem to be fairly well established that they're just using x264, anyone got any idea how they plan on making money from this, er, product/service they're offering?
[+] [-] midvar|13 years ago|reply
Video encoders are notorious for their bloated claims and faulty comparisons. After a quick perusal of their site I see they compare themselves only to web-m, and use generalized %'s to claim differences between products and processes.
They say they've optimized for human perceptual quality. Perhaps they have. But did they really do a better job than the open source H.264 alternative (x264) used by everyone from YouTube and Hulu, to Facebook? One thats been in development for _years_ now and beats just about everything thrown at it in a fair comparison? I think not. I'd love to be wrong though.
When they do eventually come out with comparisons, I hope they have read this first : http://x264dev.multimedia.cx/archives/472
[+] [-] gmartres|13 years ago|reply
So their encoder is based on x264 (without credit). The settings are slightly different between the clips, so maybe they have some magic thing to detect which settings to use. In any case, it'd be interesting to reencode the original clip using plain x264, once using the same options they used and once using some sane x264 preset.
[+] [-] Daiz|13 years ago|reply
I'm up for that. It'll take a while to download all the source videos, but I'll report back with results and videos when I'm done.
I already did a test encode with x264 --preset veryslow --tune film --level 4.0 --crf 21 on the Bourne Identity clip, and the result is within the same realm than the Beamr video with ~58% the filesize. Which is not surprising at all, considering Beamr's settings.
EDIT: Testing with nearly identical settings (x264 --bframes 0 -- subme 8 --no-psy --no-mbtree) using the latest x264 gives me about 40% the bitrate of the Beamr example for the Bourne Identity clip, which unsurprisingly looks notably worse. So maybe Beamr's "magical" encoding technology is all about fiddling with x264's CRF to make it double the bitrate at a given value compared to vanilla x264?
[+] [-] Tuna-Fish|13 years ago|reply
Just to note, no credit is necessarily required. x264 is now dual-licensed under GPL2 + commercial license. If they bought a commercial license, there's nothing illegal or wrong in what they do.
[+] [-] josephlord|13 years ago|reply
Any encoder manufacturer can find some great looking and easy to encode images that can be massively reduced in data rate without much visible impact. However in tougher scenarios they will be unlikely to beat the existing market leaders by more than 10% or so under similar constraints (latentcy/CPU/multi-pass possible) under the review of experts. Tough scenarios will be when there is lots of noise and non-uniform movement. Pop concerts can be very challenging with crowd movements, flashing lights and low general light levels leading to noise.
[+] [-] Tloewald|13 years ago|reply
Their first claimed success was with JPEGmini — this is a lot more credible since JPEG is a pretty obsolete format that simply continues because it's so well supported. JPEG2000 delivers the purported benefits of JPEGmini for real -- it's just poorly supported.
I suspect all they're really doing is fine-tuning compression settings based on differencing input and output using a decent perceptual algorithm.
I uploaded a photo straight out of a Nikon V1 onto JPEGmini and got a 2.2x reduction. I could do better than that by dialing down quality until I saw a difference. JPEG fine is way overkill for most purposes, and that's their market.
[+] [-] Daiz|13 years ago|reply
They don't seem to have done anything new or substantial by themselves, and as such, this thing is basically nothing but pure marketing bullshit. If you want to do high quality H.264 encodes, just use x264 directly.
[+] [-] unknown|13 years ago|reply
[deleted]
[+] [-] huhtenberg|13 years ago|reply
Then I took the high quality image and started bumping up standard JPEG compression level until I reduced the file size to 1.3M. That mapped to 17/100 setting, with 1 being highest quality, 100 - highest compression).
Then I made an arithmetic per-pixel diff between the original high quality image and two 1.3M versions. Cropped them to 2000 x 2000 and saved as PNGs (cropping is to fit the imgur file size limits). Here they are -
http://imgur.com/1FOwXhq
http://imgur.com/3efRooj
Now, have a look and tell me that JPEGMini hasn't got something exciting going on.
--
[0] http://media.jpegmini.com/homepageImages/20121203/hipnshoot....
[+] [-] Xcelerate|13 years ago|reply
[+] [-] conroe64|13 years ago|reply
-rw-r--r-- 1 tim users 4930482 Dec 1 21:22 hipnshoot.jpg
-rw-r--r-- 1 tim users 1310430 Feb 26 08:27 hipnshoot_gimp.jpg
-rw-r--r-- 1 tim users 1352882 Dec 3 09:16 hipnshoot_mini.jpg
I put them here: http://www.rareventure.com/jpegmini/
"hipnshoot_mini.jpg" is the "jpegmini" version and "hipnshoot_gimp.jpg" is the one I compressed with plain old gimp.
[+] [-] deadairspace|13 years ago|reply
[+] [-] mnicole|13 years ago|reply
[+] [-] ricardobeat|13 years ago|reply
[+] [-] mistercow|13 years ago|reply
That's just a marketing hype decsription of how all lossy compression works.
[+] [-] Xcelerate|13 years ago|reply
[+] [-] Hari_Seldon|13 years ago|reply
>Elements of the recordings that are easily perceived are represented with exacting precision, while other parts that are not very audible can be represented less accurately. Meanwhile, inaudible information can be discarded altogether.
Fraunhofer Institute
[+] [-] lucian1900|13 years ago|reply
[+] [-] pornel|13 years ago|reply
The way you encode lossy formats makes a lot of difference. For example the same x264 encoder gives dramatically worse results when aiming for optimal PSNR rather than using its psychovisual optimizer:
• http://x264.nl/developers/Dark_Shikari/imagecoding/x264.png
• http://x264.nl/developers/Dark_Shikari/imagecoding/x264_psnr...
http://x264dev.multimedia.cx/archives/541
[+] [-] lutusp|13 years ago|reply
I say this because claims like this tend to be very source-dependent. Change the source, change the claim.
[+] [-] Tuna-Fish|13 years ago|reply
However, this also means that if you want better encoding, you can just grab x264 and do it yourself.
[+] [-] kierank|13 years ago|reply
[+] [-] batterseapower|13 years ago|reply
[+] [-] jackau|13 years ago|reply
[+] [-] runn1ng|13 years ago|reply
Apart from the strong PR claims, this is actually not that bad, no? If you have two lossy compressions, but one is perceptually better, you will rather use the perceptually better one, no?
[+] [-] beagle3|13 years ago|reply
x264 is awesome. super awesome, even.
[+] [-] ZeroGravitas|13 years ago|reply
[+] [-] judofyr|13 years ago|reply
[+] [-] LatvjuAvs|13 years ago|reply
But when you do, yeah, I understand why size drops is that big.