0

I am working on an approach to archive our website (dynamically generated) periodically (say every month) and keep it versionned so that I can go back and pull a page at a certain period.

My initial approach is to crawl the site recursively and commit it to a subversion repository so that I can use subversions history and export feature.

Is there any other optimal solution which uses less space as possible? Also, I am not sure how long a subversion commit of an entire site would take so a faster solution is also desired.

Hennes
  • 65,804
  • 7
  • 115
  • 169

1 Answers1

1

Use 7zip on a cron job to periodically zip with recursive and a datetime stamp - then test and mv to a fail-over cluster - always test your backup periodically or else you will have corrupt data and not find out until its critical. I believe you can use a 7zip test switch as well. We've been using this approach for 7 years with mission critical data and zip 16 times a day with no failures whatsoever.

jonsca
  • 4,084