(no title)
crishoj | 1 year ago
> The particular calculation in question is to produce a random distribution. The result of this calculation has no practical use. > > They use this particular problem because it has been formally proven (with some technical caveats) that the calculation is difficult to do on a conventional computer (because it uses a lot of entanglement). That also allows them to say things like "this would have taken a septillion years on a conventional computer" etc. > > It's exactly the same calculation that they did in 2019 on a ca 50 qubit chip. In case you didn't follow that, Google's 2019 quantum supremacy claim was questioned by IBM pretty much as soon as the claim was made and a few years later a group said they did it on a conventional computer in a similar time.
l33tman|1 year ago
The RCS is a common benchmark with no practical value, as is stated several times in the blog announcement as well. It's used because if a quantum computer can't do that, it can't do any other calculation either.
The main contribution here seems to be what they indeed put first, which is the error correction scaling.
Closi|1 year ago
She doesn't even say that this isn't a big leap (she says it's very impressive - just not the sort of leap that means that there are now practical applications for quantum computers, and that a pinch of salt is required on the claim of comparisons to a conventional computer due to the 2019 paper with a similar benchmark).
pixelsort|1 year ago
This was a fascinating watch, and not the kind of content that is easy to find. Besides videos like that one, I enjoy her videos as fun way to absorb critical takes on interesting science news.
Maybe she is controversial for being active and opinionated on social media, but we need more science influencers and educators like her, who don't just repeat the news without offering us context and interpretation.
falleng0d|1 year ago
And I can't blame her for adopting this trend, in many cases it is the difference between surviving or not on YouTube nowadays.
zipy124|1 year ago
cowl|1 year ago
> Of course, as happened after we announced the first beyond-classical computation in 2019, we expect classical computers to keep improving on this benchmark
As IBM showed their estimate of classical computer time is taken out of their a**es.
perching_aix|1 year ago
Problems that benefit from quantum computing as far as I'm aware have their own formal language class, so it's also not like you have to consider Sabine's or any other person's thoughts and feelings on the subject - it is formally demonstrated that such problems exist.
Whether the real world applications arrive or not, you can speculate for yourself. You really don't need to borrow the equally unsubstantiated opinion of someone else.
eigenket|1 year ago
On the other hand it's actually not completely necessary to have a superpolynomial quantum advantage in order to have some quantum advantage. A quantum computer running in quadratic time is still (probably) more useful than a classical computer running in O(n^100) time, even though they're both technically polynomial. An example of this is classical algorithms for simulating quantum circuits with bounded error whose runtime is like n^(1/eps) where eps is the error. If you pick eps=0.01 you've got a technically polynomial runtime classical algorithm but it's runtime is gonna be n^100, which is likely very large.
vctrnk|1 year ago
> The next challenge for the field is to demonstrate a first "useful, beyond-classical" computation on today's quantum chips that is relevant to a real-world application. We’re optimistic that the Willow generation of chips can help us achieve this goal. So far, there have been two separate types of experiments. On the one hand, we’ve run the RCS benchmark, which measures performance against classical computers but has no known real-world applications. On the other hand, we’ve done scientifically interesting simulations of quantum systems, which have led to new scientific discoveries but are still within the reach of classical computers. Our goal is to do both at the same time — to step into the realm of algorithms that are beyond the reach of classical computers and that are useful for real-world, commercially relevant problems.
chvid|1 year ago
Does anyone know?