You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
Add 2NT's blogging platform to fc2-blog (same backend software as FC2)
|3 weeks ago|
|bot||1 year ago|
|cogs||2 years ago|
|config||9 years ago|
|dashboard||7 months ago|
|db||3 weeks ago|
|doc||1 year ago|
|lib||3 years ago|
|ops||2 years ago|
|pipeline||7 months ago|
|plumbing||3 years ago|
|spec||3 years ago|
|test||3 years ago|
|uploader||2 years ago|
|viewer||7 years ago|
|.gitignore||5 years ago|
|.gitmodules||9 years ago|
|.travis.yml||2 years ago|
|Gemfile||2 years ago|
|Gemfile.lock||2 years ago|
|INSTALL.backend||3 years ago|
|INSTALL.pipeline||3 years ago|
|LICENSE||9 years ago|
|README||4 years ago|
|README.pipeline-recovery||2 years ago|
|Rakefile||8 years ago|
<SketchCow> Coders, I have a question.
<SketchCow> Or, a request, etc.
<SketchCow> I spent some time with xmc discussing something we could
do to make things easier around here.
<SketchCow> What we came up with is a trigger for a bot, which can
be triggered by people with ops.
<SketchCow> You tell it a website. It crawls it. WARC. Uploads it to
<SketchCow> I can supply machine as needed.
<SketchCow> Obviously there's some sanitation issues, and it is root
all the way down or nothing.
<SketchCow> I think that would help a lot for smaller sites
<SketchCow> Sites where it's 100 pages or 1000 pages even, pretty
<SketchCow> And just being able to go "bot, get a sanity dump"
2. More info
ArchiveBot has two major backend components: the control node, which
runs the IRC interface and bookkeeping programs, and the crawlers, which
do all the Web crawling. ArchiveBot users communicate with ArchiveBot
by issuing commands in an IRC channel.
User's guide: http://archivebot.readthedocs.org/en/latest/
Control node installation guide: INSTALL.backend
Crawler installation guide: INSTALL.pipeline
3. Local use
ArchiveBot was originally written as a set of separate programs for
deployment on a server. This means it has a poor distribution story.
However, Ivan Kozik (@ivan) has taken the ArchiveBot pipeline,
dashboard, ignores, and control system and created a package intended for
personal use. You can find it at https://github.com/ArchiveTeam/grab-site.
Copyright 2013 David Yip; made available under the MIT license. See
LICENSE for details.
Thanks to Alard (@alard), who added WARC generation and Lua scripting to
GNU Wget. Wget+lua was the first web crawler used by ArchiveBot.
Thanks to Christopher Foo (@chfoo) for wpull, ArchiveBot's current web
Thanks to Ivan Kozik (@ivan) for maintaining ignore patterns and
tracking down performance problems at scale.
Other thanks go to the following projects:
* Celluloid <http://celluloid.io/>
* Cinch <https://github.com/cinchrb/cinch/>
* CouchDB <http://couchdb.apache.org/>
* Ember.js <http://emberjs.com/>
* Redis <http://redis.io/>
* Seesaw <https://github.com/ArchiveTeam/seesaw-kit>
6. Special thanks
Dragonette, Barnaby Bright, Vienna Teng, NONONO.
The memory hole of the Web has gone too far.
Don't look down, never look away; ArchiveBot's like the wind.