开发者

Using wget and cron to download webpages

开发者 https://www.devze.com 2023-01-25 09:53 出处:网络
Ok, so I know i can use: wget -r <web开发者_C百科site> > <file> to get a webpage and save it. My question is, how would i use cron and wget to get a webpage on an hourly, or even min

Ok, so I know i can use:

wget -r <web开发者_C百科site> > <file>

to get a webpage and save it. My question is, how would i use cron and wget to get a webpage on an hourly, or even minute basis, and then save them into a folder, zip and tarball it, and then keep adding to it for a review at a later date.

I know i can manually do this, my goal is to basically download it ever 10- 20 minutes, for roughly 4 hours (doesn't matter if it goes longer) and append the all into a nice directory, then zip said directory to conserve space, and check them later in the day.


To edit cron table

crontab -e

You can add an entry like this

0,20,40 * * * *  wget URL ~/files/file-`date > '+%m%d%y%H%M'`.html &

To download/save the file every 20 mins.

Here it is a small reference about crontab expressions so you can adjust the values

To TAR the files automatically the crontab would be slightly complex:

0,20,40 * * * *  wget URL > ~/files`date '+%m%d%y'`/file-`date '+%H%M'`.html &
* 12 * * *       tar cvf ~/archive-`date '+%m%d%y'`.tar ~/files`date '+%m%d%y'`

This would do it at noon, if you want to do it at mifnight it's more complex because you need to TAR the previous day but I think with this you'll get the idea.


Or without cron:

for i in `seq 1 10`; do wget -r http://google.de -P $(date +%k_%M) && sleep 600; done

10 times, every 10 minutes

EDIT: Use zip like this

zip foo.zip file1 file2 allfile*.html
0

精彩评论

暂无评论...
验证码 换一张
取 消

关注公众号