A website that irrevocably deletes itself once indexed by Google.
The site is constantly searching for itself in Google, over and over and over, 24 hours a day. The instant it finds itself in Google search results, the site will instantaneously and irrevocably securely delete itself. Visitors can contribute to the public content of the site, these contributions will also be destroyed when the site deletes itself.
Why would you do such a thing? My full explanation was in the content of the site.
(edit: …which is now gone)
UPDATE: The experiment lasted 22 days before it was indexed by Google on 24 February 2015 at 21:01:14 and instantaneously destroyed. It was primarily shared via physical means in the real world, word of mouth, etc.
If you want to conduct your own similar experiment, the source code is here.
Nothing has been done to prevent the site from being indexed, however the
NOARCHIVE meta tag is specified which instructs Google not to cache
their own copy of the content.
The content for this site is stored in memory only (via Redis) and is loaded
in via a file from an encrypted partition on my personal laptop. This
partition is then destroyed immediately after launching the site. Redis
backups are disabled. The content is flushed from memory once the site
detects that it has been indexed.
The URL of the site can be algorithmically generated and is configured via
environment variable, so this source code can be made public without
disclosing the location of the site to bots.
Visitors can leave comments on the site while it is active. These comments
are similarly flushed along with the rest of the content upon index event,
making them equally ephemeral.
Sample configuration notes for running on Heroku:
$ heroku create `pwgen -AnB 6 1` # generates a random hostname $ heroku addons:add rediscloud # default free tier disables backups $ heroku config:set REDIS_URL=`heroku config:get REDISCLOUD_URL` $ heroku config:set SITE_URL=`heroku domains | sed -ne "2,2p;2q"` $ git push heroku master $ heroku run npm run reset $ heroku addons:add scheduler:standard $ heroku addons:open scheduler
Schedule a task every N minutes for
npm run-script query (unfortunately seems
like this can only be done via web interface).
scripts/load_content.js to load the content piped from STDIN.
You can configure monitoring to check the
/status endpoint for
"OK" if you
trust an external service with your URL.
Original URL: http://feedproxy.google.com/~r/feedsapi/BwPx/~3/LkGT1vGEC8k/README.md