You'll have to developp a tool using php5.3 and MySQL5.
No warning, no Notice will appear (fix them, don't hide them).
You'll use mysqli php extension (and not mysql extension).
This tool will scan Awstats html files, generated by [url removed, login to view], and copy them into html files.
Tool will be used into command line, and script parameter will be:
- url of awstats
- path where you'll put local copy of awstats
Goal is to get a backup of [url removed, login to view] layout into html files.
So take care, you'll have to rewrite all urls in order to make them work everywhere directory will be putted.
And take care too, to awstats frame.
Of course, you won't have to copy external pages.
Awstats urls are looking like this:
[url removed, login to view];year=2009&config=myconfig&framename=mainright&output=lasthosts
In order to work with huge volume of data, spider should walks along only updated pages.
To do that, you'll have to write all scanned into MYSQL Database.
url : is scanned url (VARCHAR 255)
ended : true/false (BOOL)
Take a look to GET parameters:
if month+year < NOW : put true to 'ended' field, and don't scan again this page because it won't change anymore.
else : put false to 'ended' field, and you'll have to scan it again when script will be reloaded.
A detail: local copies of files should have in the top this tag: <!-- AUTOBACKUP -->