top | item 17195692

The Never-Ending War on Fake Reviews

121 points| raleighm | 7 years ago |newyorker.com | reply

100 comments

order
[+] endtime|7 years ago|reply
Disclaimer: Not speaking for my employer, and my experience was several years ago and may be out of date.

My first project at Google was working on fake reviews on Maps. ML was certainly useful for some things, but at the end of the day sometimes you literally have no practical way of knowing if a review is fake or not. Was this written by a customer or by the business owner's brother in law? Or by his competitor down the street? Who knows? Which means it's pretty hard to get good, complete training data to train your model on.

Of course, there are classes of fake reviews that are easy enough to detect. But as the fake review-writing AIs get better, I don't know how the anti-fake review AIs can win.

[+] dmortin|7 years ago|reply
> at the end of the day sometimes you literally have no practical way of knowing if a review is fake or not

Can't you check the account itself? If the account's behavior over time shows signs it's an actual human user instead of a bot.

This can't filter the owner's brother in law, but if bot accounts can be filtered then they cannot leave lots of reviews either, because the owner does not have hundreds of relatives.

[+] sametmax|7 years ago|reply
Well, if a human can't tell, at some point, the AI can't tell either.

Since humans are good at creating lies, and that we all get fooled at some moment in our life, it means the problem hasn't been solved for millennial and that it's not an AI problem.

Sure you can throw at the AI way more data than we used to have, but even totalitarian state didn't manage to silenced the opposition so I doubt a business bound by law can do anything about it.

Google protection measures are already becoming super invasive and instead of helping me, has locked me out of my account several times because it detected I was a fraud.

There is a limit on what you can do properly here.

[+] jimkleiber|7 years ago|reply
I'm starting to think that the solution is not to reduce fake reviews, but to increase verified accounts who give reviews.

I think in increasing the number of reviews from verified accounts (this account verified to be connected with human of name X), the rest of the anonymous or pseudonymous reviews may be looked at with more scrutiny.

Right now, let's assume that the default for social networks and other places of interaction on the internet is something like 10% verified accounts and 90% unverified. We know that within that 90% unverified, there are many real humans with their real names. So we spend a lot of effort to parse that 90% to find the real humans, the ones pretending to be others, the bots, the anonymous ones, and others. Companies are using ML, individuals like me just guess and try to only accept people that I've met in person and avoid platforms where bots seem to hang. However, with sites like Amazon, I have to hope that Amazon has filtered out the reviews from the fake accounts. Yes, there will be fake reviews from verified accounts, and that seems like another issue, and maybe less prevalent, I'm not sure.

If the percentage of verified accounts flips, from 10% verified / 90% unverified to 90% verified / 10% unverified, I think many of these fake reviews filter themselves out. We would trust that this specific person is saying this, and still leave room for pseudonyms and anonymity, but be more clear that one is using such measures to hide their identity.

How do you think this would impact fake reviews on Amazon? What secondary impacts do you see more verified accounts on Amazon and on other platforms having on discourse? I hope this post wasn't too long, again, I'm new to HN and hope I'm staying within bounds.

[+] bostik|7 years ago|reply
Even for verified purchases, Amazon's over-eager review beggary is just bonkers.

You buy something, and regardless of what it is, you get automated requests to review the purchase. Sometimes before the item has arrived. Practically always well before you have managed to establish whether the purchase was any good or not.

This is why I ignore pretty much every single please-review-me prod from Amazon and their retailers. Send me a request to review before I could have had proper time to evaluate the purchase in practice, and you get bucketed with other entitled f--kwits. You are also likely to lose my future business.

I believe Amazon could improve the S/N ratio of their reviews if they actually considered how long it takes to test out any of their purchases.

[+] ballenf|7 years ago|reply
I don't think it addresses the core issue and creates additional ones. There are people who have valuable opinions who won't post them if connected back to them. There are people who won't bother getting verified who would otherwise leave honest reviews.

The bad players will find a way to get verified and now you're worse off than before (fewer legit reviews plus bad ones with a fraudulent stamp of approval).

Whatever the solution you have to take into account that bad players have a higher incentive to pass your test / jump through hoops than legitimate people do.

[+] solotronics|7 years ago|reply
I think this would help to solve some of the quality and fake goods issues on Amazon. What it can't solve is a dishonorable seller switching out a good product for an inferior one. We should have 'verified sellers' as well tied to a real name.
[+] nitwit005|7 years ago|reply
Amazon and some of the other firms have enough data to know a subset of their users are real people without any additional effort. I have years of order history, with changing credit cards and addresses tied to my name, plus books and movie viewing over a long period. It'd never be worth the effort to fake that.
[+] ljm|7 years ago|reply
I actually wonder why crowd-sourced reviews have become so ubiquitous, to the point of almost being necessary (along with five-star ratings). Has anyone legitimately attempted to challenge the status quo here?

I'm reminded of videogamedunkey on YouTube and his video about game critics[0], where he goes into some detail about the preferences and integrity of the critic being just as important (if not more) than the review and rating.

Professional critics build a strong reputation around their taste and preference and they gain notoriety from doggedly sticking to those principles. So you'd likely trust the opinion of Roger Ebert or Mark Kermode if you share their taste in cinema, and even watch something you normally wouldn't if even they recommended it. You're very likely to subscribe to a publication whose critics align with your own tastes because they're effectively curating content for you in the form of recommendations.

None of that applies when you have reviews from a succession of total strangers - you're not going to research dozens or hundreds of commentators to establish a logical consistency in their point of view and decide whether or not their tastes align with yours. More often than not the reviews are low quality and low value, spread across a five-point scale but essentially treated as a binary like/didn't like system.

At that point, crowd-sourced reviews tell you nothing you didn't know already: some people enjoyed it for one reason and others didn't for a different reason. How do you know if those reasons are legit or authentic when they're not fake? How do you know if they weren't gamed or incentivised somehow to inflate expectations? Which ones can you trust?

I suppose they boost sales but just like advertising, that doesn't mean they're automatically beneficial to the consumer. It's just another attempt at manipulation.

[0] https://www.youtube.com/watch?v=lG2dXobAXLI

[+] ohashi|7 years ago|reply
I think one of the biggest challenges is scale/breadth. For some things like Movies, we have lots of critics and there is demand for it. I work in web hosting reviews. Everyone technical has an opinion and non-technical too. The average HN reader and a non technical person starting their first blog are very different use cases and demands. I see way too often the 'why don't you just get a VPS and do it yourself it's cheaper' advice to basically anyone. Those are the kinds of people quite actively trying to 'help' newbies. And the reviews, well, it's not a product most people try a lot of different companies nor have any sense of how theirs might stack up. I've got to be one of the most experienced people in terms of number of companies used (I probably had accounts at ~50 companies this year), and I still wouldn't be comfortable writing reviews. It's too subjective and biased.
[+] sien|7 years ago|reply
Metacritic does this for games, music, tv and film.

Aggregating what they regard as reliable reviewers is pretty reasonable.

For general products it's much harder. There is less of a market. But perhaps it'll come. Aggregate Ars Technica, Tom's Hardware and some other hardware review sites maybe?

[+] Analemma_|7 years ago|reply
I'm starting to despair of the possibility of ever having a reliable way to gauge product quality on the Internet. It seems to be a huge asymmetry problem where, as soon as a trustworthy system appears, the rewards of gaming it are so big compared to what consumers are willing to pay to have it stay trustworthy, that sellers will throw unlimited money at the problem until they have captured it with useless garbage.

About the only thing you can do is find a particular writer or site that has a reputation you trust, but they might never review the specific product you're interested in. And when you're trying to buy something in an entirely new category you're not familiar with, it's hopeless.

[+] mnm1|7 years ago|reply
This is why it's important to have a reliable return process for online shops. Unfortunately, no online store that I know of has one, least of all Amazon, so I find myself buying more and more in regular stores and generally buying less and less often simply to reduce the risk of getting stuck with a bad purchase. It's ironic that the online shopping experience now has been topped by the brick and mortar store experience without brick and mortar doing anything significant to alter their experience in decades. Online stores are simply untrustworthy and I don't see this changing soon. From Amazon actively closing accounts they don't like to them and others failing to ship products repeatedly, the experience is now worse in every single way. And I'm not a big fan of driving to stores at all.
[+] supertrope|7 years ago|reply
Goodhart's law: A measure ceases to be useful as soon as it becomes an optimization target.

Concentrated benefit and diffuse cost is how a minority can maintain a globally sub-optimal status quo.

[+] 3pt14159|7 years ago|reply
All we need is a web of trust that is easy to use and so many of these problems go away.
[+] jvz|7 years ago|reply
Fake reviews can be seen as an instance of Goodhart's law, where the metric is the rating or score of the business. Initially those ratings may have high correlation with something real, let's say the "quality" of the business. But the more people rely on those scores and the reviews underlying them, the more incentive businesses have to game the system— which destroys the original correlation between ratings and quality.

A big part of the problem with review systems is the one-to-many nature of nearly all of them: when a person posts a review, that review and its score can be seen by everyone. This leverage makes it very efficient for businesses to game the system, as a small amount of fake information can "infect" the purchasing decisions of a large number of users.

So, one alternative might be a many-to-many review system where you only see reviews and ratings from your network of friends/follows (and maybe friends-of-friends, to increase coverage). So essentially Twitter, but with tools and UI that focus on reviews and ratings. That way, fake reviews could only affect a limited number of people, making the cost/benefit calculus much less attractive for would-be astroturfers and shills.

[+] tankerslay|7 years ago|reply
Doesn't this already violate a bunch of laws? (undisclosed advertising, wire fraud or something like it, etc) Just enforce them. Can't do much about overseas people but you can penalize the businesses that purchase fraudulent marketing. E.g. if doctors started getting medical licenses permanently revoked for buying fake "testimonials" I imagine it would end fairly quickly.
[+] dfraser992|7 years ago|reply
https://douglas-fraser.com/FakeReviews/index.html

I did my dissertation on this topic - text only analysis. I used a dataset that was commonly used in the beginning, but there are some issues with it. I plan to extend this to real reviews, as in 80 million Amazon ones (when I get the time).

Text based features are useful, but non-text based ones are even more so. Even spamming groups can be detected; at least there has been research into that. Combining all the techniques in an ensemble would be productive - but is it really in Amazon's interest? My sense is whatever they do, they pick the low hanging fruit and trying to process every review that comes in would require a lot of CPUs perhaps. But stuff like floods of reviews for new products that are fairly similar should be easy to detect. Perhaps they are relying on Fakespot and reviewmeta to do the heavy lifting.

[+] greggman|7 years ago|reply
curious if all the fake reviews will end up creating a market for certified reviewers or something like consumer reports. Sure we can't trust cnet or wired or pcmag but maybe if the incentive is high enough there will be a market for honest reviews including an incentive for auditing and verification?

could start a reviewer guild and use crypto signatures to verify their guild membership is up to date and still valid. hoping the guild has the incentive to stay honest etc.

[+] rogerbinns|7 years ago|reply
> creating a market for certified reviewers or something like consumer reports.

I'd argue that is exactly what The Wirecutter is. https://thewirecutter.com/ https://en.wikipedia.org/wiki/Wirecutter_(website)

They solve the reviewer problem by having in house staff write the recommendations (reviews). Those staff then find and bring in experts. Virtually every page has a "Why you should trust us" section listing experience of those who contributed to the article. It does mean you get credible opinions, but also that you don't get the "wisdom of the masses" such as at Amazon.

[+] supertrope|7 years ago|reply
There was always a problem with misaligned incentives. If you're not paying for the review, they have no loyalty to the reader.

e-commerce websites pay lip service to deleting fake reviews but higher product review scores result in more sales. Even if this doesn't result in intentionally ignoring fake reviews in pursuit of short term sales growth, note how Amazon.com like all other five star rating systems suffers from massive score inflation. Most have a 4.5/5.0 rating. 4.0 indicate some potential problems (or less sophisticated customers), and 3.5 means it's sub-par.

A lot of online stores simply have no customer review section because they have rationally determined that for them decreased conversions due to bad reviews and moderation costs exceed increase in conversions due to good reviews.

[+] plankers|7 years ago|reply
The guild could issue ReviewCoin and use this to pay reviewers, thus incentivizing them to work in good faith lest the public lose faith in the project, tanking the value of the coin.

ReviewCoin could be redeemed at participating online retailers for goods, thus allowing reviewers to buy more products to review.

This is a silly idea but it makes me smile.

[+] otakucode|7 years ago|reply
Of course if the reviews you saw were only from people whose prior reviews you agreed with or who posted reviews which matched up with the reviews you yourself posted, then the system just wouldn't be game-able. But that's involve getting users actively involved in writing reviews as well as actively agreeing with the views of others. But when it comes down to it, until you have a review system that will tell me that the Super Bowl is garbage and not worth watching because it knows my tastes and doesn't give a shit about what other people that are nothing like me think, your review system is a piece of garbage. Of course it's gamed. It's trash, and it's not going to somehow be magically not-trash even if people don't actively seek to game it.
[+] joering2|7 years ago|reply
I think the bigger problem is reviews written by stiupid people. I seen a mosquito killer lamp that had very low reviews - when you show only one star, majority reviewers were complaining that the blue light lamp doesnt kill roaches and it inly attracts mosquitos, so it must be fake.
[+] supertrope|7 years ago|reply
I prefer the term insufficiently savvy customers. Products that are not universally compatible or required specific actions to be taken tend to review poorer. Cheap items tend to review better as customers are more forgiving.

The customer ultimately determines their satisfaction level, even if they are complaining that a software product clearly described as Windows only is incompatible with their Mac. Or their "wireless" machine doesn't work without the power cord inserted.

[+] chrisfinne|7 years ago|reply
I use fakespot.com regularly for Amazon purchases.
[+] jurassic|7 years ago|reply
I think video reviews will become more and more important. Sure, people can lie on camera, but I think it’s harder to fake convincingly. And AIs/spammers can’t present on camera with the same ease that they throw up blog posts.
[+] luddaite|7 years ago|reply
This reminds of an article I read many years ago about the Facebook's and Google's approaches to organizing information. The article argued that Facebook seeks to organize data in the context of human relationships. I wonder if the future of online reviewing will be one where people on trust reviews left by people they know. Facebook supports leaving reviews on company pages (AFAIK) but overall it seems like they haven't put much effort behind making a robust reviewing system. I can easily see them coming to dominate this market though.
[+] supertrope|7 years ago|reply
This is the logic behind calling someone to verify who they are. Harder to fake with automated means, easy for people to verify if they know the person.

The potential downside is that people can sometimes become overly trusting of truer to life higher bandwidth mediums. e.g. Fake IRS, grandson needs money for bail and baby (cue crying noise).

[+] lordnacho|7 years ago|reply
This area of the internet seems poised for a massive disruption as AI starts coming in. In fact I'd be a little surprised if it wasn't already a somewhat influential thing.

You could imagine a GAN-like ecosystem appearing, where you have loads of algorithms both writing and identifying fake reviews.

And with that, poor old humans like myself can surely not compete. You'd always have to have a healthy scepticism towards reviews. I already do, in an era where most of the fakes are probably (?) still human-authored.

[+] erikb|7 years ago|reply
Has anybody considered that fake reviews might be part of the system, not part of a problem? Of course it's not in our interests as customers, but then again I have never paid anybody for a fair review, but I bet a lot of wealthy review targets are willing to pay a lot for a clean public appearance. Would a review collector or shopping site really be willing to not help with that?
[+] bsparker|7 years ago|reply
I'm starting a new project for verified reviews: reviews that are "peer reviewed" by others in the community. I am still in stealth mode but I'm working to solve this: https://thepeerreview.com
[+] ballenf|7 years ago|reply
If it's good and covers products of interest to me, I'd consider paying a moderate subscription fee for such a service. But only if that meant the site didn't have advertising, tracking or affiliate links.

The power of affiliate links to warp reviews is underestimated even on Wirecutter, imo. If they have to choose between a product that has no affiliate links and one that does, it's pretty much impossible for that not to eventually affect the recommendations.

[+] Cilvic|7 years ago|reply
I like this idea, but how to you suggest dealing with fake peers reviewing the fake reviews?
[+] politician|7 years ago|reply
How do you deal with collusion rings?
[+] deegles|7 years ago|reply
One solution I just thought of (a bit scorched earth maybe) would be to stop allowing written reviews. Only allow video reviews and enforce showing a) the reviewers face b) the product and c) the proof of purchase (could be blurred afterwards for privacy). An automatically created transcript would be shown with the video.

This would make it more difficult/obvious if one person were to submit many reviews (use face recognition), raise the barrier for fake reviews, and give a lot more ‘signal’ to people trying to determine if a review is fake.

Of course, with Deep Fakes and such this could be bypassed still, but it could still have an impact.

[+] function_seven|7 years ago|reply
The problem there is you pretty much get no reviews at that point. Almost nobody is going to go through all that trouble to leave a review of a product if it means having to film yourself doing so. Meanwhile, paid reviewers will happily do so.
[+] InitialLastName|7 years ago|reply
This is still easy to get around.

1) Give a person a gift card to purchase product

2) They purchase product and review (following this procedure)

3) Pay them

4) Repeat

[+] lopmotr|7 years ago|reply
It's pretty common for clothing buyers on Taobao to post a photo of themselves wearing what they bought as a review. I'm not sure how the site or the sellers incentivise this, but it surely adds some trust.
[+] specialist|7 years ago|reply
Detecting fake news, reviews, astroturfing, fraud after the fact is too late, futile, confusing.

Correct answer is to have verifiable sources, citations.

What we used to call "journalism".

[+] wyattpeak|7 years ago|reply
That's all well and lovely, but the reason reviews have always been popular is that in many domains people trust the man on the street more than they do a journalist or reviewer, for often perfectly valid reasons.

If a restaurant site removed all user reviews and replaced them with a food critic's opinions I would trust it less, not more, fake reviews notwithstanding.

[+] KhayriRRW|7 years ago|reply
Facebook page reviews is a good example of this.

~ Khayri R.R. Woulfe

[+] throwawaymath|7 years ago|reply
Hey I don't mean to be a jerk, but please don't sign your name on every comment you write. Your username is right there, if I want to know who you are I can just look up a centimeter from your comment.
[+] Nasrudith|7 years ago|reply
The term itself is problematic - haven't we learned from "fake news" that said accurate description of literally whatever gets the most clicks will rapidly be turned into a rhetorical cudgel that means "news that I don't like"?

Given the sheer potential for abuse I would advise great caution with measures to remedy. This problem is ancient given that it has been around for literal centuries at least. Just in the article a contemporary of Oscar Wilde used the technique!

[+] BookmarkSaver|7 years ago|reply
I don't think there's any comparison to be made. The term "fake news" was primarily popularized by a guy who was blatantly lying and who's grasp of the English language barely extends to 5 letter words and who doesn't even really bother with the grammar. "Fake reviews" are literally just fake reviews, it's not even really a technical term, it's just what they are.
[+] supertrope|7 years ago|reply
Fake reviews are like spam and troublemakers at a comedy club. Nobody but the assholes themselves likes it. Clubs that don't throw out hecklers or fist fighters will fail when people go elsewhere that doesn't suck. There is a clear definition of non-genuine reviews: people who have never tried the product, they are being paid for the review, or you are the seller posting under a sock puppet account.