shell脚本----周期压缩备份日志文件

一、日志文件样式


二、目标


1、备份压缩.log结尾&&时间样式为“date +%Y%m%d”的日志文件(如:20170912、20160311等)

2、可指定压缩范围(N天前至当天):如:今天、昨天(date -d "-1 day" +%Y%m%d)至今天、前天至今天

压缩命名格式为:日期.tar.gz(或:日期.tar.bz2),压缩N天范围内文件后将会生成N个压缩包

3、可指定压缩模式(二选一):tar czf  或  tar cjf

4、可指定删除范围:删除N天前的日志文件,如:今天为20170912,删除3天前日志,即删除20170908及其以前的日志文件

5、每周五(不一定周五,可在脚本中指定周几打包)将压缩文件打包(打包成功后删除压缩文件),并上传到日志服务器,上传成功后删除打包文件

6、默认备份当天日志文件,压缩模式为“czf”,删除三天前日志文件

三、脚本内容


#!/bin/bash
#author:xiami
#date:20170907
#description: compress files of a specified mode and delete logs before a particular date
#version:v0.1

strategy_logs_path="/root/apps/myapp/log"
date=$(date +%Y%m%d)


#------------初始化函数默认参数----------
init_argv(){  #example:init_argv -d 4 -m cjf -r 5 打包压缩近四天及当天日志,压缩模式cjf,删除五天前日志
    compress_date=$(date +%Y%m%d)
    backup_mode="czf"
    del_days_num=3
    backup_path="/root/apps/logs_backup"
    line=$(ps -ef |grep "ssh-agent"|awk '{if($0!~/grep ssh-agent/)print $0}'|wc -l)
    while getopts d:m:r: opt
    do
        case $opt in
        d) compress_days="$OPTARG" ;;
        m) backup_mode="$OPTARG" ;;
        r) del_days_num="$OPTARG" ;;
       ?) echo "Usage: `basename $0` [d|m|r]"
           echo "-d 'Integer' (tar & compress Integer day logfile)"
           echo "-m 'czf|cjf' (mode)"
           echo "-r 'Interger' (remove files Integer days ago)"
           exit 1 ;;
        esac
    done
    if [ ! -d "$backup_path" ];then
        mkdir -p $backup_path
    fi
#    if [ line -lt 1 ];then
#        echo "Not running ssh-agent"
#        exit 1
#    fi
}
#------------打包并压缩日志文件操作函数--------------------
tar_compress_log(){    #参数1:压缩日期  参数2:压缩备份模式
    local compress_date="$1"
    local backup_mode="$2"
    cd "$strategy_logs_path"
#    ls . -name "*$date*.log"|xargs tar $backup_mode "$date.tar.gz"   #压缩当天日志
    if [ "$backup_mode" == 'czf' ];then
        ls | grep ".*[0-9]{8}.*.log$"|grep ".*"$compress_date".*"|xargs tar $backup_mode "$backup_path/"$1".tar.gz"
    elif [ "$backup_mode" == 'cjf' ];then
        ls | grep ".*[0-9]{8}.*.log$"|grep ".*"$compress_date".*"|xargs tar $backup_mode "$backup_path/"$1".tar.bz2"
    else
        echo "tar mode error";exit 2
    fi
}
#------------打包并压缩N天前到当前日期范围的日志文件----------
compress_range_date(){     #参数1:压缩备份天数 参数2:压缩备份模式
    local dates="$date"
    local compress_days="$1"
    local backup_mode="$2"
    if [[ ! -z "$compress_days" && "$compress_days" -ne 0 ]];then
        for i in `seq 1 "$compress_days"`;do
               dates=$(date -d "-$i day" +%Y%m%d)
            #   n=$(echo $dates|awk '{print NF}')
               tar_compress_log "$dates" "$backup_mode"
        done
    else
        tar_compress_log "$dates" "$backup_mode"
    fi
}
#------------每周五打包当周压缩文件并在打包成功后删除单个压缩文件------------
tar_file(){
    weekday=$(date +%u)
    if [ "$weekday" -eq 5 ];then
        cd "$backup_path"
        ls|xargs tar cf "$date-Fri-logs.tar" && rm *.tar.?z*
    fi
}
#-----------删除(del_log_days)天前日志文件(根据mtime)---------
delete_days_log1(){   #参数1:del_days_num
    local del_days_num="$1"
    cd "$strategy_logs_path"
    find . -type f -mtime +$del_days_num |grep ".*[0-9]{8}.*.log"|xargs rm aa -rf  #删除操作,请谨慎
}
#----------删除(del_log_days)天前日志文件(根据log文件名)----
delete_days_log2(){    #参数1:del_days_num
    local del_days_num="$1"
    cd "$strategy_logs_path"
    local num=$(ls |grep ".*[0-9]{8}.*.log$" |grep -o "[0-9]{8}"|awk '!day[$0]++'|wc -l)
    local files=$(ls |grep ".*[0-9]{8}.*.log$" |grep -o "[0-9]{8}"|awk '!day[$0]++')
    if [ ! -z "$num" ];then
        local field=""
        for i in  `seq 1 "$num"`
        do
            field=$(echo $files|awk '{print $v}' v=$i)
            if [ "$field" -lt $(date -d "-$del_days_num day" +%Y%m%d) ];then
                    ls *$field* |grep   '.*[0-9]{8}.*.log$'|xargs rm aa -rf
            fi
        done
    fi
}
#-----------上传到日志服务器-------------
upload_tgz(){
    scp $backup_path/$date.tar.gz 192.168.119.133:/opt/ && rm $backup_path/$date.tar.gz
   #scp $backup_path/$date*-logs.tar 192.168.119.133:/opt/ && rm $backup_path/$date*-logs.tar
}
#-----------main-------------------------
main(){
    init_argv $argv
    compress_range_date "$compress_days" "$backup_mode"
    delete_days_log2 $del_days_num
    tar_file
#    upload_tgz
}
#----------执行部分----------------------
argv=""
until [ "$#" -eq 0 ];do
            argv="$argv $1"
            shift
done
main $argv
log_tar_clean.sh

四、运行脚本


日志文件初始数量为图一所示

4.1 带参数运行后

4.2  不带参数运行(默认情况,备份当天日志文件,压缩模式为“czf”,删除三天前日志文件)

 为了使测试环境一致,首先恢复被删除日志文件,删除压缩文件

不带参数运行脚本

 五、计划任务



脚本完善:

1、删除每天的日志文件,机器中只保留近N天(默认14天)的压缩文件

2、考虑到安全性,文件上传改为:远端日志服务器从机器拉取每天对应的日志文件,而不是机器主动上传。

3、一次备份多个目录下日志,目录列表存放于scan_logdir.inc

完善后脚本:

#!/bin/bash
#author:xiami
#date:20170907
#description: compress files of a specified mode and delete logs before a particular date, reference: http://www.cnblogs.com/xiami-xm/p/7511087.html
#version:v0.2

date=$(date +%Y%m%d)
scan_log_dir="$(dirname $0)/scan_logdir.inc"

#------------初始化函数默认参数----------
init_argv(){  #example:init_argv -d 4 -m cjf -r 5 打包压缩近四天及当天日志,压缩模式cjf,删除五天前日志
    compress_date=$(date +%Y%m%d)
    backup_mode="cjf"
    del_days_num=3
    backup_path="/root/apps/logs_backup"
    stay_days=14
    line=$(ps -ef |grep "ssh-agent"|awk '{if($0!~/grep ssh-agent/)print $0}'|wc -l)
    while getopts d:m:r:s: opt
    do
        case $opt in
        d) compress_days="$OPTARG" ;;
        m) backup_mode="$OPTARG" ;;
        r) del_days_num="$OPTARG" ;;
        s) stay_days="$OPTARG" ;;
       ?) echo "Usage: `basename $0` [d|m|r]"
           echo "-d 'Integer' (tar & compress Integer day logfile)"
           echo "-m 'czf|cjf' (mode)"
           echo "-r 'Interger' (remove files Integer days ago)"
           exit 1 ;;
        esac
    done
    if [ ! -d "$backup_path" ];then
        mkdir -p $backup_path
    fi
}
#------------(实际操作函数)打包并压缩日志文件,压缩成功后立即将文件删除--------------------
tar_compress_log(){    #参数1:压缩日期  参数2:压缩备份模式
    local compress_date="$1"
    local backup_mode="$2"
    local logs_path="$3"
    cd "$logs_path"
#    cd "$strategy_logs_path"
#    ls . -name "*$date*.log"|xargs tar $backup_mode "$date.tar.gz"   #压缩当天日志
    if [ "$backup_mode" == 'czf' ];then
        ls | grep ".*[0-9]{8}.*.(log|csv)$"|grep ".*"$compress_date".*"|xargs tar $backup_mode $backup_path/dir$a.$compress_date.tar.gz >/dev/null 2>&1 && ls | grep ".*[0-9]{8}.*.(log|csv)$"|grep ".*"$compress_date".*"|xargs rm -rf
    elif [ "$backup_mode" == 'cjf' ];then
        ls | grep ".*[0-9]{8}.*.(log|csv)$"|grep ".*"$compress_date".*"|xargs tar $backup_mode $backup_path/dir$a.$compress_date.tar.bz2 >/dev/null 2>&1 && ls | grep ".*[0-9]{8}.*.(log|csv)$"|grep ".*"$compress_date".*"|xargs rm -rf
    else
        echo "tar mode error";exit 2
    fi
}
#------------检查scan_logdir.inc是否有数据-------------------
check_scan(){
    local line=$(cat "$scan_log_dir" | wc -l)
    if [ "$line" -eq 0 ];then
        exit 3;
    fi
}
#------------需要备份的日志目录-------------------------------
backup_all_logs_dir(){
    check_scan
    local a=1
    while read line;do
        local logs_path="$line"
        compress_range_date "$compress_days" "$backup_mode" "$logs_path"
#        delete_days_log1 "$del_days_num" "$logs_path"
        delete_days_log2 "$del_days_num" "$logs_path"
        let a++;echo "-----------------$a----------------------------------"
    done < "$scan_log_dir"
}
#------------打包并压缩N天前到当前日期范围的日志文件----------
compress_range_date(){     #参数1:压缩备份天数 参数2:压缩备份模式
    local dates="$date"
    local compress_days="$1"
    local backup_mode="$2"
    local logs_path="$3"
#    if [[ ! -z "$compress_days" && "$compress_days" -ne 0 ]];then
    if [ ! -z "$compress_days" ];then
        for i in `seq 0 "$compress_days"`;do
               dates=$(date -d "-$i day" +%Y%m%d)
            #   n=$(echo $dates|awk '{print NF}')
               tar_compress_log "$dates" "$backup_mode" "$logs_path"
        done
    else
        tar_compress_log "$dates" "$backup_mode" "$logs_path"
    fi
}
#------------每周五打包当周压缩文件并在打包成功后删除单个压缩文件------------
tar_file(){
    weekday=$(date +%u)
    if [ "$weekday" -eq 5 ];then
        cd "$backup_path"
        ls|xargs tar cf "$date-Fri-logs.tar" && rm *.tar.?z*
    fi
}
#-----------只保留14天内的压缩日志文件------------------------------------
stay_tgz_log(){
    local stay_days="$stay_days"
    local rm_date=$(date -d "-$stay_days day" +%Y%m%d)
    cd $backup_path
    ls |grep -o "[0-9]{8}"|awk '!aa[$0]++'|while read line;do
        if [ "$line" -le $rm_date ];then
             ls *$line* |xargs rm -rf
        fi
    done

}
#-----------删除(del_log_days)天前日志文件(根据mtime)---------
delete_days_log1(){   #参数1:del_days_num
    local del_days_num="$1"
    local logs_path="$2"
#    cd "$strategy_logs_path"
    cd "$logs_path"
    find . -type f -mtime +$del_days_num |grep ".*[0-9]{8}.*.(log|csv)$"|xargs rm -rf  #删除操作,请谨慎
}
#----------删除(del_log_days)天前日志文件(根据log文件名)----
delete_days_log2(){    #参数1:del_days_num
    local del_days_num="$1"
    local logs_path="$2"
#    cd "$strategy_logs_path"
    cd "$logs_path"
    local num=$(ls |grep ".*[0-9]{8}.*.(log|csv)$" |grep -o "[0-9]{8}"|awk '!day[$0]++'|wc -l)
    local files=$(ls |grep ".*[0-9]{8}.*.(log|csv)$" |grep -o "[0-9]{8}"|awk '!day[$0]++')
    if [ ! -z "$num" ];then
        local field=""
        for i in  `seq 1 "$num"`
        do
            field=$(echo $files|awk '{print $v}' v=$i)
            if [ "$field" -le $(date -d "-$del_days_num day" +%Y%m%d) ];then
                    ls *$field* |grep   '.*[0-9]{8}.*.(log|csv)$'|xargs rm -rf
            fi
        done
    fi
}
#-----------上传到日志服务器-------------
upload_tgz(){
    scp $backup_path/$date.tar.gz 192.168.119.133:/opt/ && rm $backup_path/$date.tar.gz
   #scp $backup_path/$date*-logs.tar 192.168.119.133:/opt/ && rm $backup_path/$date*-logs.tar
}
#-----------main-------------------------
main(){
    init_argv $argv
    backup_all_logs_dir
#    compress_range_date "$compress_days" "$backup_mode"
#    delete_days_log2 $del_days_num
    stay_tgz_log $stay_days
#    tar_file
#    upload_tgz
}
#----------执行部分----------------------
argv=""
until [ "$#" -eq 0 ];do
            argv="$argv $1"
            shift
done
main $argv
clean_log.sh

日志服务器拉取脚本:

#!/bin/bash
date=$(date +%Y%m%d)
remote_logs_backup=/root/apps/logs_backup
local_logs_backup=/root/apps/local_logs_backup
hostsfile="$(dirname $0)/hosts.inc"
while read line;do
    if [ ! -d "$local_logs_backup/$line" ];then
        mkdir -p "$local_logs_backup/$line"
    fi
    scp root@$line:$remote_logs_backup/*$date* $local_logs_backup/$line >/dev/null 2>&1
#&& ssh -nl root $line "rm $remote_logs_backup/* -rf" >/dev/null 2>&1   #ssh登陆或拷贝操作时若文件不存在会报错,可忽略
done < $hostsfile
pull_logs.sh
[root@xiamihost3 script]# cat hosts.inc 
192.168.119.131   #hosts.inc中是目标机器的ip地址
192.168.119.133

  

基本功能已实现,但脚本还有待改进,例如脚本中使用的grep可以换成egrep,可以少一层转义。

上传备份功能还未进行测试,可能会出现错误,当然,上传前肯定得先将本地公钥放入服务器的authorized_keys文件中,实现ssh无密登陆。

最后想感慨一句,还是python更强大

good good study,day day up

原文地址:https://www.cnblogs.com/xiami-xm/p/7511087.html