There is no explicit per-user limit.
There are some performance limits derived from the limits of the hardware: for example, the number of running browsers is limited and the queue size is limited.
They are very rarely hit, but the user willing to submit tons of urls has good chance to hit them.
1. same pages like in your example are very rare. ads or block of recomedations or feed of tweets differ.
2. it might take considerable time to make a hidden snapshot to compare the previous snapshot with; also this operation may fail due to network conditions or server error.
3. for some usercases it has sense to same exact the same page twice, to proof that it has not changed.
Images are deduplicated, htmls are not. There are too many images which are the same across thousands of snapshots; for example, the icons of the social networks
Could you help the free project and improve the translation?
Alternatively, you can change the language: http://www.w3.org/International/questions/qa-lang-priorities.en.php#changing
I plan to index the whole archive using elasticsearch and use it also for url-search.
No exact date yet.
Presumably, never.
My view on the trust issue is different: it is better to sacrifice the global availability a bit (it is far from being 100% anyway) than to censor the content.
Not yet.
You can try archive.org, webcitation.org or scribd.com.
Few hours ago, an user found XSS vulnerability on both archive.org and archive.is.
Page https://archive.is/VSGzW saved from https://archive.org/search.php?query=1XSS&sort=-publicdate<svg%20onload=confirm(/XSSPOSED/)> contained executable javascript.
The bug is fixed.
It seems I just created a good page to buy the F-Secure ad-space :)
There is also Opera Turbo, for free.
In case if the number of abuses increases it may end up in pre-moderated submissions from Tor .
So far they are post-moderated with special attention to ones made from Tor IPs.