In the near future, our team plans to launch a unique service that combines the capabilities of the Internet Archive (archive.org) and a search engine.
We plan to index the main pages of all sites that have ever been saved in the web archive. Our website database will also contain an archive of various metrics such as Alexa, Ahrefs, Majestic, keywords, WHOIS domain and other historical data from the entire lifetime of the website. Thus, it will be possible to search for the desired site to restore drops or extract deleted content by a huge number of parameters, such as keywords in text and tags, traffic on a specific date, the presence of historical links, name server, and so on. Using this system, it will be possible to make selections of any complexity, for example, find all domains deleted in 2018 containing the words "webmaster analytics", which then had an Alexa rating of less than 300k and which had more than 30 unique visitors per day for a specific keyword in 2016 year. Our service will have a convenient interface for finding the required data and will contain screenshots of websites with all parameters in the form of graphs and tables.
The next stage in the development of the system will be the indexing of live sites and the expansion of the indexed content base. For search, media files and some internal pages of the site will be available, selected according to an algorithm that takes into account the importance of this page. As a result, the system will use mainly its own database of archived sites and become independent of Archive.org.
The use of article materials is allowed only if the link to the source is posted: https://archivarix.com/en/blog/archivarix-net/
Sometimes our users ask why the website was not fully restored? Why the website doesn't it work the way I would like it to? Known issues when restoring sites from archive.org.…
Two big tasty coupons are valid from Friday 27.11.2020 to Monday 30.11.2020. Each of them gives a balance bonus in the form of 20% or 50% of the amount of your last or new payment.…
Three years ago, on September 29, 2017, our archive.org downloader service was launched. All these 3 years we have been continuously developing, we have created our own CMS, a Wordpress plugin, a syst…
Wayback Machine ( web.archive.org ) Alternative. Internet archive search engine. Find archived copies of websites. Data from 1996. Full-text search.
In the near future, our team plans to launch a uni…
How to generate meta name="description" on all pages of a website? How to make the site work not from the root, but from a subdirectory?…
How to show hidden files on macOS. How to view and edit files starting with dot ( like .htaccess ) in macOS?…
Our Website downloader system allows you to download up to 200 files from a website for free. If there are more files on the site and you need all of them, then you can pay for this service. Download …
This article describes regular expressions used to search and replace content in websites restored using the Archivarix System. They are not unique to this system. If you know the regular expressions …
In order to make it convenient for you to edit the websites restored in our system, we have developed a simple Flat File CMS consisting of just one small php file. Despite its size, this CMS is a powe…
The Wayback Machine is the famous and biggest archive of websites in the world. It has more than 400 billion pages on their servers. Is there any archiving services like Archive.org? …