I had a client once who had something similar, although unintentionally. She approached me because her website "kept getting hacked" and she didn't trust the original developers to solve the security problems... And rightly so!
There were two factors that, together, made this happen: first, the admin login form was implemented in JS, and if you went to log in with it with JS disabled, it wouldn't verify your credentials. And it submitted via a GET request. Second, once you were in the admin interface, you could delete content from the site by clicking on an X in the CMS. Which, as was the pattern, presented you with a JS alert() prompt before deleting the content... via a GET request.
Looking at the server logs around the time it got "hacked", you could see GoogleBot happily following all the delete links in the admin interface.
> I had a client once who had something similar, although unintentionally.
I did that too. I was aware of the problem, but at the time (1996) I did not know how to fix it.
So I just documented it and warned that they should keep the site away from altavista.
This was back before cookies had wide support, so login state was in the URL. If you allowed a search spider to know that URL it would have deleted the entire site by spidering it.
I did eventually fix it by switching to forms, and strengthening the URL token to expire if unused for a while. And then eventually switching to cookies (at one point it supported both url tokens and cookies).
I have not thought about those days in such a long time.
I accidentally deleted about half of the database at a startup where I’d recently started working by approximately the same method. I was running a copy of the web interface on my laptop, connecting over the internet to our MySQL server, and also running ht://dig’s spider on localhost from cron. It started spidering the delete links. Fortunately, I’d also started running daily MySQL backups from cron (there were no backups before I started working there), so we only lost a few hours of everyone’s work. As you can imagine, though, they weren’t super happy with me that day.
I agree that there are lots of reasons that someone would make a site like this, but I think people are curious as to the maker's specific reason. From the github:
Why would you do such a thing? My full explanation was in the content of the site. (edit: ...which is now gone)
I'm curious as to what the website said originally.
I think more and more the word hacker has lost its original meaning at least in this community. If I were reading a similar story on a tor hidden service, let's say, I would not be asking why, but here I do.
An alternative would be to check for the browser user agent and delete the website right at that point and return a 404 page to the Google crawler bot. Then Google won't have a static copy of the website.
Your approach is "a website that irrevocably deletes itself once indexed by Google".
What OP has done is "a website that irrevocably deletes itself once Google decided to publicly reveal the fact that it indexed said website".
OP's approach has no way of knowing when the site was indexed. It's conceivable that Google indexed it on the very first day and decided not to share it publicly until 21 days later.
What about the opposite? A website that created when it is indexed? Start with nothing and content is added each time the site is visited by Googlebot, or shared on Facebook, tweeted, posted on Reddit, etc. The website exists only so that it can be shared, and the act of sharing it defines what the website is.
This is an uber cool idea. Especially if, when this website is shared by someone, it would attempt to scan the sharer's public feed, last submissions, last comments, last tweets, etc. (depending on where it got shared), and generate additional content based on what it found.
Postmodernism is a lot more relevant to the digital age than anything, imo. It emphasizes pointing out ways of thinking and doing, which I think is especially relevant when we are actually automating most of our ways of thinking and doing.
I know it gets a bad rap because of the ridiculous examples, but the real point of it engages the viewer into a serious kind of contemplation concerning the massive infrastructure that exists and how that shapes our culture, thoughts, understanding, action..
We have the expectation that the generations to come will accept this infrastructure and what it says about how the human mind functions. But much of it is founded on belief systems of how thought and action operate in the real world. Most of these systems are baseless, the idea of a base obfuscated only by the sheer complexity involved in understanding each layer.
As far as I can tell, you just posted part of a random screengrab from your web browser for no obvious reason. Striking's response suggests that this is actually a reference to a site which, per the OP, is gone forever, along with any chance of getting your joke. So...I'm not really sure what you were expecting.
I see some potential use of this, for example as soon as Google crawlers reach the site I know that it is accessible from outside and I destroy the site.
[+] [-] tonyarkles|11 years ago|reply
There were two factors that, together, made this happen: first, the admin login form was implemented in JS, and if you went to log in with it with JS disabled, it wouldn't verify your credentials. And it submitted via a GET request. Second, once you were in the admin interface, you could delete content from the site by clicking on an X in the CMS. Which, as was the pattern, presented you with a JS alert() prompt before deleting the content... via a GET request.
Looking at the server logs around the time it got "hacked", you could see GoogleBot happily following all the delete links in the admin interface.
[+] [-] ars|11 years ago|reply
I did that too. I was aware of the problem, but at the time (1996) I did not know how to fix it.
So I just documented it and warned that they should keep the site away from altavista.
This was back before cookies had wide support, so login state was in the URL. If you allowed a search spider to know that URL it would have deleted the entire site by spidering it.
I did eventually fix it by switching to forms, and strengthening the URL token to expire if unused for a while. And then eventually switching to cookies (at one point it supported both url tokens and cookies).
I have not thought about those days in such a long time.
[+] [-] ikeboy|11 years ago|reply
[+] [-] kragen|11 years ago|reply
[+] [-] toxicFork|11 years ago|reply
[+] [-] tlrobinson|11 years ago|reply
Hackers don't need a reason, other than it being clever, novel, fun, etc. But if you want a reason there are plenty:
* art: there are numerous interpretations of this
* fun: this is sort of the digital equivalent of a "useless box" http://www.thinkgeek.com/product/ef0b/
* science: experiment to see how widespread a URL can be shared without Google becoming aware of it
* security: embed unique tokens in your content to detect if it has leaked to the public
[+] [-] barbs|11 years ago|reply
Why would you do such a thing? My full explanation was in the content of the site. (edit: ...which is now gone)
I'm curious as to what the website said originally.
[+] [-] soheil|11 years ago|reply
[+] [-] RexRollman|11 years ago|reply
[+] [-] raimondious|11 years ago|reply
[+] [-] dsjoerg|11 years ago|reply
[+] [-] LukeB_UK|11 years ago|reply
[+] [-] unknown|11 years ago|reply
[deleted]
[+] [-] rjempson|11 years ago|reply
[+] [-] frik|11 years ago|reply
[+] [-] desdiv|11 years ago|reply
What OP has done is "a website that irrevocably deletes itself once Google decided to publicly reveal the fact that it indexed said website".
OP's approach has no way of knowing when the site was indexed. It's conceivable that Google indexed it on the very first day and decided not to share it publicly until 21 days later.
[+] [-] LukeB_UK|11 years ago|reply
[+] [-] ForHackernews|11 years ago|reply
> the NOARCHIVE meta tag is specified which prevents the Googles from caching their own copy of the content.
[+] [-] whoopdedo|11 years ago|reply
[+] [-] TeMPOraL|11 years ago|reply
Sounds like an awesome weekend project.
[+] [-] yk|11 years ago|reply
[+] [-] cheatsheet|11 years ago|reply
Postmodernism is a lot more relevant to the digital age than anything, imo. It emphasizes pointing out ways of thinking and doing, which I think is especially relevant when we are actually automating most of our ways of thinking and doing.
I know it gets a bad rap because of the ridiculous examples, but the real point of it engages the viewer into a serious kind of contemplation concerning the massive infrastructure that exists and how that shapes our culture, thoughts, understanding, action..
We have the expectation that the generations to come will accept this infrastructure and what it says about how the human mind functions. But much of it is founded on belief systems of how thought and action operate in the real world. Most of these systems are baseless, the idea of a base obfuscated only by the sheer complexity involved in understanding each layer.
[+] [-] egypturnash|11 years ago|reply
If that ain't baroque I don't know what is.
[+] [-] psykovsky|11 years ago|reply
[+] [-] ssalazar|11 years ago|reply
[+] [-] byte1918|11 years ago|reply
http://i.imgur.com/cjDeLEb.png
EDIT: What's with the downvote hate? Somebody actually posted a valid key...
[+] [-] PhasmaFelis|11 years ago|reply
[+] [-] striking|11 years ago|reply
[+] [-] hackhat|11 years ago|reply
So anyone really understood why he did this?
[+] [-] TeMPOraL|11 years ago|reply
[+] [-] ikeboy|11 years ago|reply
[+] [-] TimWolla|11 years ago|reply
[+] [-] imjustsaying|11 years ago|reply
[+] [-] WA|11 years ago|reply
- deleted after 100 visitors
- deleted if visited with IE 6.0 for the first time
- deleted if referrer is Facebook
- ...
[+] [-] comboy|11 years ago|reply
[+] [-] cubano|11 years ago|reply
[+] [-] thewizardofmys|11 years ago|reply
[+] [-] placeybordeaux|11 years ago|reply
[+] [-] aqme28|11 years ago|reply
[+] [-] arash_milani|11 years ago|reply
[+] [-] ars|11 years ago|reply
[+] [-] neilellis|11 years ago|reply
[+] [-] scottcanoni|11 years ago|reply
1. Sending the NOINDEX meta tag
2. Combining meta tags
3. Monitoring for a referrer URL that matches a Google search page to catch the 1st non-sneaky user coming from the index.
4. Monitoring other search engines and their behaviors.
[+] [-] angelortega|11 years ago|reply
[+] [-] shubhamjain|11 years ago|reply
[+] [-] tjgq|11 years ago|reply
[+] [-] hartator|11 years ago|reply
[+] [-] bernardlunn|11 years ago|reply
[+] [-] lukasm|11 years ago|reply