Show HN: DNS over Wikipedia
398 points| aaronjanse | 5 years ago |github.com
I saw a thread a while ago (linked in README) discussing how Wikipedia does a good job keeping track of the domains of websites like Sci-Hub or The Pirate Bay. Someone mentioned checking Wikipedia to find links to these sites, so I thought this would be a fun thing to automate!
To try it out, install an extension or modify your hosts file, then type in the name of a website with the TLD `.idk`.
For example: scihub.idk -> sci-hub.tw
Cheers!
_gjrn|5 years ago
If an analogy was needed with a network service perhaps this is more like a proxy redirector than DNS.
Keep in mind: with this you will still be misdirected if your DNS/hosts file is pointing the name into a different IP than it should be.
capableweb|5 years ago
> Resolve DNS queries using the official link found on a topic's Wikipedia page
@aaronjanse: you probably want to correct this. "Resolving DNS records" carry a specific meaning in that you have a DNS record and you "resolve" it to a value, which actually. You're kind of doing, in a way, I suppose.
I was convinced when I started writing this comment that calling this "resolve dns queries" is wrong. But thinking about it, DNS resolving is not necessarily resolving a "name into a IP-address" as @HugoDaniel in the comment I'm replying to is saying (think CNAME records and all the others that don't have IP addresses). It's just taking something and making it into something else, traditionally over DNS servers. But I guess you could argue that this is resolving a name into a different name, that then gets resolved into a IP address. So it's like a overlay over DNS resolving.
Meh, in the end I'm torn. Anyone else wanna give it a shot?
Polylactic_acid|5 years ago
basch|5 years ago
mathieubordere|5 years ago
frei|5 years ago
itaysk|5 years ago
bausano_michael|5 years ago
hk__2|5 years ago
iakh|5 years ago
nitrogen|5 years ago
nicbou|5 years ago
cpach|5 years ago
segfaultbuserr|5 years ago
cxr|5 years ago
> Not as trivially compromised as it sounds like it would be; could be faked with (inevitably short-lived) edits, but temporality can't be faked. If a system were rolled out tomorrow, nothing that happens after rollout [...] would alter the fact that for the last N years, Wikipedia has understood that the website for Facebook is facebook.com. Newly created, low-traffic articles and short-lived edits would fail the trust threshold. After rollout, there would be increased attention to make sure that longstanding edits getting in that misrepresent the link between domain and identity [can never reach maturity]. Would-be attackers would be discouraged to the point of not even trying.
https://www.colbyrussell.com/2019/05/15/may-integration.html...
Asmod4n|5 years ago
BillinghamJ|5 years ago
pishpash|5 years ago
hk__2|5 years ago
This should be Wikidata. Wikipedia does that, but this is more and more moved into Wikidata. This is a good thing, because Wikidata is much easier to query, and the official website of an entity is stored at a single place, that is then reused by all articles about that entity in all languages.
snek|5 years ago
The rust server set up with dnsmasq is a legit DNS server though.
MatthewWilkes|5 years ago
Vinnl|5 years ago
Just as a heads-up of what you could expect to see happening :)
abiogenesis|5 years ago
usmannk|5 years ago
stepanhruda|5 years ago
renewiltord|5 years ago
LinuxBender|5 years ago
- Domains used by the site (first party)
- Domains used by the site (third party)
- Methods allowed per domain.
- CDN's used by the site
- A records and their current IP addresses
- Reporting URL for errors
Then include the public keys for that payload in DNS and in the APEX of the domain? Perhaps a browser add-on could verify the content and report errors back to a standard reporting URL with some technical data that would show which ISP is potentially being tampered with? Does something like this already exist beyond DANE? Similar to HSTS maybe the browser could cache some of this info and show diffs in the report? Maybe the crypto keys learned for a domain could also be cached and warn the user if something has changed (show diff and option to report)? Maybe more complex would be a system that allows a consensus aggregation of data to be ingested by users so they may start off in a hostile network and some trusted domains populated by the browser in advance, also similar to HSTS?
andrekorol|5 years ago
blattimwind|5 years ago
For example looking up "sci hub" on Wikidata leads to https://www.wikidata.org/wiki/Q21980377 which has an "official website" field.
oefrha|5 years ago
CapriciousCptl|5 years ago
1. Look at past wiki edits combined with article popularity or other signals to arrive at something like a confidence level.
2. Offer some sort of confirmation check to the user.
gbear605|5 years ago
[1]: https://en.wikipedia.org/w/index.php?title=Equifax&diff=9455...
edjrage|5 years ago
Edit: Unrelated to this issue, but I have a more general idea for the kinds of inputs this extension may accept. It could be an omnibox command [0] that takes the input text, passes it through some search engine with "site:wikipedia.org", visits the first result and finally grabs the URL. So you don't have to know any part of the URL - you can just type the name of the thing.
[0]: https://developer.chrome.com/extensions/omnibox
yreg|5 years ago
29athrowaway|5 years ago
jrockway|5 years ago
Overall this is a nifty hack and I like it a lot. Wikipedia has an edit history, and a DNS changelog is something that is very interesting to have. People can change things and phish users of this service, of course, but with the edit log you can see when and potentially why. That kind of transparency is pretty scary to someone that wants to do something malicious or nefarious.
jhasse|5 years ago
leoh|5 years ago
O_H_E|5 years ago
erikig|5 years ago
- How do you handle ambiguity? e.g what happens when sci-hub.idk and scihub.idk differ?
- Aren’t you concerned by the fact that Wikipedia is open to editing by the public?
aaronjanse|5 years ago
Arguably the thrill of uncertainty could add to the fun :D
captn3m0|5 years ago
tubbs|5 years ago
I used Pushbullet's recipe for "Google acquisitions" up until the night I got the notification "Google acquires 4chan". After being perplexed for a bit and a few more "acquisitions" were made, I discovered the recipe just used Wikipedia's List of mergers and acquisitions by Alphabet[1] page as a source.
[1]: https://en.wikipedia.org/wiki/List_of_mergers_and_acquisitio...
jneplokh|5 years ago
Regardless, having a system where you can base it off a website could definitely be expanded beyond Wikipedia. Great work!
snorrah|5 years ago
upgoat|5 years ago
jaimex2|5 years ago
newswasboring|5 years ago
Sabinus|5 years ago
sm4rk0|5 years ago
https://lmddgtfy.net/?q=%5Chacker%20news
That's especially useful if DDG is default search engine in your browser.
(I'm not affiliated with DDG)
kelnos|5 years ago
The purpose of this seems to be to treat Wikipedia as a trusted, reliable source of truth about the canonical URL for websites (debatable, of course). The idea is that you don't trust the search engines, perhaps because you live in a country where your government has required search engines to censor results in some way, but (for some reason?) lets you go to Wikipedia.
rootsudo|5 years ago
unknown|5 years ago
[deleted]
jakear|5 years ago
How interesting. Bing doesn't do this, which leads me to believe it's not a matter of legality. Is Google simply electing to self-censor results that it'd prefer it's used not to know about? Strange move, especially given the alternative Google does index is almost definitely more nefarious.
sixhobbits|5 years ago
As many have pointed out, this just makes it easier for actually malicious sites to get traffic.
[0] https://torrentfreak.com/google-downranks-65000-pirate-sites...
tomcooks|5 years ago
At the bottom of the page click on the DMCA complaint, you'll find all the URLs you shouldn't ever, never ever, click on~
jonchurch_|5 years ago
I'm curious if that's the case for you as well, or if it's my ISP blocking (I wouldn't expect to see the cloudflare error if my ISP was blocking but I don't know).
I bring this up because if the site is unresponsive from wherever you're searching (or perhaps unresponsive for all, idk) then maybe it got de-ranked on google.
aequitas|5 years ago
[0] https://blog.iusmentis.com/2017/06/19/eu-hof-verklaart-the-p...
jimmaswell|5 years ago
BubRoss|5 years ago
lizardmancan|5 years ago
[deleted]
nomanlaghari|5 years ago
[deleted]