"When the lights go out and the signals fade,
And the silence of the void begins to wade,
This archive stands, a beacon in the night,
Preserving wisdom, keeping truth in sight.
For if the world should crumble, dust to dust,
In this digital vault, we place our trust."
WorldEndArchive is a resilient, autonomous web crawler designed to preserve human knowledge in the face of catastrophe. It tirelessly traverses the web, capturing essential information and storing it in accessible formats for a post-internet world.
Setup Gatherer (The Crawler)
The Gatherer traverses the web to collect data.
-
Navigate to directory:
cd gatherer -
Install dependencies:
npm install
-
Configure:
- Copy
.env.exampleto.env. - Edit
.envto set your proxies, concurrency, etc.
- Copy
-
Start Crawling:
npm start
- Dashboard: http://localhost:3000
- Output: Data is saved to
gatherer/output/.
Setup Client (The Reader)
The Client allows you to browse the archived content offline.
-
Navigate to directory:
cd client -
Install dependencies:
npm install
-
Start Reader:
npm start
- Interface: http://localhost:3001
Edit gatherer/.env to customize:
MAX_CONCURRENCY: Number of simultaneous requests.DELAY_BETWEEN_REQUESTS_MS: Throttle speed.USE_PROXIES: Enable/disable proxy rotation.KEYWORDS_FILE: Path to your custom keywords.
If you are reading this, the internet is likely still operational. Feel free to contribute to the codebase to ensure it is ready for when it is not.
Preserving the past, for the future.