summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
author李玺康 <[email protected]>2020-05-21 14:28:46 +0800
committer李玺康 <[email protected]>2020-05-21 14:28:46 +0800
commit1ad1dc757971d9abd3cda23560a0c18490cc8514 (patch)
tree85a7f37a985673e70648fcf9a79e66228a8f57e9
parent9fce5c40b28812998b76972d447a199ec34f2cee (diff)
首个ansible版本HEADmaster
-rw-r--r--external-flume/conf/radius/radius_k2f.properties22
-rw-r--r--external-flume/conf_external/active/active_defence_k2f.properties22
-rwxr-xr-xexternal-flume/conf_external/active/clean_start_flume.sh (renamed from external-flume/conf/connection/clean_start_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/active/count_flume.sh (renamed from internal-flume/conf/connection/count_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/active/dae_active_denfence.sh (renamed from internal-flume/conf/k2f/dae_f2k.sh)6
-rwxr-xr-xexternal-flume/conf_external/active/flume-env.sh (renamed from external-flume/conf/connection/flume-env.sh)0
-rw-r--r--external-flume/conf_external/active/restart_log/restart_connection_f2k.log (renamed from internal-flume/conf/connection/restart_log/restart_connection_f2k.log)39
-rwxr-xr-xexternal-flume/conf_external/active/start_flume.sh (renamed from external-flume/conf/connection/start_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/active/stop_flume.sh (renamed from internal-flume/conf/connection/stop_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/connection/clean_start_flume.sh (renamed from external-flume/conf/proxy/clean_start_flume.sh)0
-rw-r--r--external-flume/conf_external/connection/connection_k2f.properties (renamed from external-flume/conf/connection/connection_k2f.properties)0
-rwxr-xr-xexternal-flume/conf_external/connection/count_flume.sh (renamed from external-flume/conf/connection/count_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/connection/dae_connection.sh (renamed from external-flume/conf/connection/dae_connection.sh)0
-rwxr-xr-xexternal-flume/conf_external/connection/flume-env.sh (renamed from external-flume/conf/proxy/flume-env.sh)0
-rw-r--r--external-flume/conf_external/connection/log4j.properties (renamed from external-flume/conf/connection/log4j.properties)0
-rw-r--r--external-flume/conf_external/connection/logs/flume.log201
-rw-r--r--external-flume/conf_external/connection/pro/2f2k.properties (renamed from external-flume/conf/connection/pro/2f2k.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/4f2k.properties (renamed from external-flume/conf/connection/pro/4f2k.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/f2k.properties (renamed from external-flume/conf/connection/pro/f2k.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/f2k1.properties (renamed from external-flume/conf/connection/pro/f2k1.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/f2k2.properties (renamed from external-flume/conf/connection/pro/f2k2.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/f2k3.properties (renamed from external-flume/conf/connection/pro/f2k3.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/file.properties (renamed from external-flume/conf/connection/pro/file.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/file1.properties (renamed from external-flume/conf/connection/pro/file1.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/file2.properties (renamed from external-flume/conf/connection/pro/file2.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/k2f.properties (renamed from external-flume/conf/connection/pro/k2f.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/kafka.properties (renamed from external-flume/conf/connection/pro/kafka.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/kafka1.properties (renamed from external-flume/conf/connection/pro/kafka1.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/kafka2.properties (renamed from external-flume/conf/connection/pro/kafka2.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/log4j.properties (renamed from external-flume/conf/connection/pro/log4j.properties)0
-rw-r--r--external-flume/conf_external/connection/pro/tail.properties (renamed from external-flume/conf/connection/pro/tail.properties)0
-rw-r--r--external-flume/conf_external/connection/restart_log/restart_connection_k2f.log (renamed from external-flume/conf/radius/restart_log/restart_connection_k2f.log)4
-rw-r--r--external-flume/conf_external/connection/restart_log/restart_f2k.log (renamed from external-flume/conf/connection/restart_log/restart_f2k.log)0
-rw-r--r--external-flume/conf_external/connection/restart_log/restart_security_f2k.log (renamed from external-flume/conf/connection/restart_log/restart_security_f2k.log)0
-rwxr-xr-xexternal-flume/conf_external/connection/start_flume.sh (renamed from external-flume/conf/proxy/start_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/connection/stop_flume.sh (renamed from external-flume/conf/connection/stop_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/proxy/clean_start_flume.sh (renamed from external-flume/conf/radius/clean_start_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/proxy/count_flume.sh (renamed from external-flume/conf/proxy/count_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/proxy/dae_proxy.sh (renamed from external-flume/conf/proxy/dae_proxy.sh)0
-rwxr-xr-xexternal-flume/conf_external/proxy/flume-env.sh (renamed from external-flume/conf/radius/flume-env.sh)0
-rw-r--r--external-flume/conf_external/proxy/log4j.properties (renamed from external-flume/conf/proxy/log4j.properties)0
-rw-r--r--external-flume/conf_external/proxy/logs/flume.log359
-rw-r--r--external-flume/conf_external/proxy/pro/2f2k.properties (renamed from external-flume/conf/proxy/pro/2f2k.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/4f2k.properties (renamed from external-flume/conf/proxy/pro/4f2k.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/f2k.properties (renamed from external-flume/conf/proxy/pro/f2k.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/f2k1.properties (renamed from external-flume/conf/proxy/pro/f2k1.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/f2k2.properties (renamed from external-flume/conf/proxy/pro/f2k2.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/f2k3.properties (renamed from external-flume/conf/proxy/pro/f2k3.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/file.properties (renamed from external-flume/conf/proxy/pro/file.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/file1.properties (renamed from external-flume/conf/proxy/pro/file1.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/file2.properties (renamed from external-flume/conf/proxy/pro/file2.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/k2f.properties (renamed from external-flume/conf/proxy/pro/k2f.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/kafka.properties (renamed from external-flume/conf/proxy/pro/kafka.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/kafka1.properties (renamed from external-flume/conf/proxy/pro/kafka1.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/kafka2.properties (renamed from external-flume/conf/proxy/pro/kafka2.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/log4j.properties (renamed from external-flume/conf/proxy/pro/log4j.properties)0
-rw-r--r--external-flume/conf_external/proxy/pro/tail.properties (renamed from external-flume/conf/proxy/pro/tail.properties)0
-rw-r--r--external-flume/conf_external/proxy/proxy_k2f.properties (renamed from external-flume/conf/proxy/proxy_k2f.properties)0
-rw-r--r--external-flume/conf_external/proxy/restart_log/restart_proxy_k2f.log5
-rwxr-xr-xexternal-flume/conf_external/proxy/start_flume.sh (renamed from external-flume/conf/radius/start_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/proxy/stop_flume.sh (renamed from external-flume/conf/proxy/stop_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/security/clean_start_flume.sh (renamed from external-flume/conf/security/clean_start_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/security/count_flume.sh (renamed from external-flume/conf/radius/count_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/security/dae_security.sh (renamed from external-flume/conf/security/dae_security.sh)0
-rwxr-xr-xexternal-flume/conf_external/security/flume-env.sh (renamed from external-flume/conf/security/flume-env.sh)0
-rw-r--r--external-flume/conf_external/security/log4j.properties (renamed from external-flume/conf/radius/log4j.properties)0
-rw-r--r--external-flume/conf_external/security/logs/flume.log189
-rw-r--r--external-flume/conf_external/security/pro/2f2k.properties (renamed from external-flume/conf/radius/pro/2f2k.properties)0
-rw-r--r--external-flume/conf_external/security/pro/4f2k.properties (renamed from external-flume/conf/radius/pro/4f2k.properties)0
-rw-r--r--external-flume/conf_external/security/pro/f2k.properties (renamed from external-flume/conf/radius/pro/f2k.properties)0
-rw-r--r--external-flume/conf_external/security/pro/f2k1.properties (renamed from external-flume/conf/radius/pro/f2k1.properties)0
-rw-r--r--external-flume/conf_external/security/pro/f2k2.properties (renamed from external-flume/conf/radius/pro/f2k2.properties)0
-rw-r--r--external-flume/conf_external/security/pro/f2k3.properties (renamed from external-flume/conf/radius/pro/f2k3.properties)0
-rw-r--r--external-flume/conf_external/security/pro/file.properties (renamed from external-flume/conf/radius/pro/file.properties)0
-rw-r--r--external-flume/conf_external/security/pro/file1.properties (renamed from external-flume/conf/radius/pro/file1.properties)0
-rw-r--r--external-flume/conf_external/security/pro/file2.properties (renamed from external-flume/conf/radius/pro/file2.properties)0
-rw-r--r--external-flume/conf_external/security/pro/k2f.properties (renamed from external-flume/conf/radius/pro/k2f.properties)0
-rw-r--r--external-flume/conf_external/security/pro/kafka.properties (renamed from external-flume/conf/radius/pro/kafka.properties)0
-rw-r--r--external-flume/conf_external/security/pro/kafka1.properties (renamed from external-flume/conf/radius/pro/kafka1.properties)0
-rw-r--r--external-flume/conf_external/security/pro/kafka2.properties (renamed from external-flume/conf/radius/pro/kafka2.properties)0
-rw-r--r--external-flume/conf_external/security/pro/log4j.properties (renamed from external-flume/conf/radius/pro/log4j.properties)0
-rw-r--r--external-flume/conf_external/security/pro/tail.properties (renamed from external-flume/conf/radius/pro/tail.properties)0
-rw-r--r--external-flume/conf_external/security/restart_log/restart_connection_k2f.log (renamed from external-flume/conf/connection/restart_log/restart_connection_k2f.log)0
-rw-r--r--external-flume/conf_external/security/restart_log/restart_f2k.log (renamed from external-flume/conf/radius/restart_log/restart_f2k.log)0
-rw-r--r--external-flume/conf_external/security/restart_log/restart_security_f2k.log (renamed from external-flume/conf/radius/restart_log/restart_security_f2k.log)0
-rw-r--r--external-flume/conf_external/security/restart_log/restart_security_k2f.log4
-rw-r--r--external-flume/conf_external/security/security_k2f.properties (renamed from external-flume/conf/security/security_k2f.properties)0
-rwxr-xr-xexternal-flume/conf_external/security/start_flume.sh (renamed from external-flume/conf/security/start_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/security/stop_flume.sh (renamed from external-flume/conf/radius/stop_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/traffic/clean_start_flume.sh (renamed from internal-flume/conf/connection/clean_start_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/traffic/count_flume.sh (renamed from external-flume/conf/security/count_flume.sh)0
-rwxr-xr-xexternal-flume/conf_external/traffic/dae_connection.sh (renamed from external-flume/conf/radius/dae_radius.sh)4
-rwxr-xr-xexternal-flume/conf_external/traffic/flume-env.sh (renamed from internal-flume/conf/connection/flume-env.sh)0
-rw-r--r--external-flume/conf_external/traffic/log4j.properties (renamed from external-flume/conf/security/log4j.properties)0
-rw-r--r--external-flume/conf_external/traffic/logs/flume.log201
-rw-r--r--external-flume/conf_external/traffic/pro/2f2k.properties (renamed from external-flume/conf/security/pro/2f2k.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/4f2k.properties (renamed from external-flume/conf/security/pro/4f2k.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/f2k.properties (renamed from external-flume/conf/security/pro/f2k.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/f2k1.properties (renamed from external-flume/conf/security/pro/f2k1.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/f2k2.properties (renamed from external-flume/conf/security/pro/f2k2.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/f2k3.properties (renamed from external-flume/conf/security/pro/f2k3.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/file.properties (renamed from external-flume/conf/security/pro/file.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/file1.properties (renamed from external-flume/conf/security/pro/file1.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/file2.properties (renamed from external-flume/conf/security/pro/file2.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/k2f.properties (renamed from external-flume/conf/security/pro/k2f.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/kafka.properties (renamed from external-flume/conf/security/pro/kafka.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/kafka1.properties (renamed from external-flume/conf/security/pro/kafka1.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/kafka2.properties (renamed from external-flume/conf/security/pro/kafka2.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/log4j.properties (renamed from external-flume/conf/security/pro/log4j.properties)0
-rw-r--r--external-flume/conf_external/traffic/pro/tail.properties (renamed from external-flume/conf/security/pro/tail.properties)0
-rw-r--r--external-flume/conf_external/traffic/restart_log/restart_connection_k2f.log (renamed from external-flume/conf/security/restart_log/restart_connection_k2f.log)4
-rw-r--r--external-flume/conf_external/traffic/restart_log/restart_f2k.log (renamed from external-flume/conf/security/restart_log/restart_f2k.log)0
-rw-r--r--external-flume/conf_external/traffic/restart_log/restart_security_f2k.log (renamed from external-flume/conf/security/restart_log/restart_security_f2k.log)0
-rwxr-xr-xexternal-flume/conf_external/traffic/start_flume.sh5
-rwxr-xr-xexternal-flume/conf_external/traffic/stop_flume.sh (renamed from external-flume/conf/security/stop_flume.sh)0
-rw-r--r--external-flume/conf_external/traffic/traffic_k2f.properties22
-rw-r--r--external-flume/flume/file/flume.tarbin0 -> 291342420 bytes
-rw-r--r--external-flume/flume/tasks/load_flume.yml115
-rw-r--r--external-flume/flume/tasks/main.yml1
-rw-r--r--external-flume/flume/templates/docker-compose.yml.j224
-rw-r--r--external-flume/flume/templates/k2fActDef.properties.j275
-rw-r--r--external-flume/flume/templates/k2fConnRecord.properties.j276
-rw-r--r--external-flume/flume/templates/k2fProxyEvent.properties.j275
-rw-r--r--external-flume/flume/templates/k2fSecEvent.properties.j275
-rw-r--r--external-flume/flume/templates/k2fTraffic.properties.j222
-rw-r--r--external-flume/flume/vars/main.yml25
-rw-r--r--internal-flume/ansible_role/flume/file/flume.tarbin0 -> 23381 bytes
-rw-r--r--internal-flume/ansible_role/flume/tasks/load_flume.yml66
-rw-r--r--internal-flume/ansible_role/flume/tasks/main.yml1
-rw-r--r--internal-flume/ansible_role/flume/templates/docker-compose.yml.j240
-rw-r--r--internal-flume/ansible_role/flume/vars/main.yml39
-rw-r--r--internal-flume/conf/k2f/k2f.properties22
-rw-r--r--internal-flume/conf/k2f/pro/2f2k.properties24
-rw-r--r--internal-flume/conf/k2f/pro/4f2k.properties58
-rw-r--r--internal-flume/conf/k2f/pro/f2k.properties24
-rw-r--r--internal-flume/conf/k2f/pro/f2k1.properties23
-rw-r--r--internal-flume/conf/k2f/pro/f2k2.properties23
-rw-r--r--internal-flume/conf/k2f/pro/f2k3.properties23
-rw-r--r--internal-flume/conf/k2f/pro/file.properties29
-rw-r--r--internal-flume/conf/k2f/pro/file1.properties29
-rw-r--r--internal-flume/conf/k2f/pro/file2.properties29
-rw-r--r--internal-flume/conf/k2f/pro/k2f.properties22
-rw-r--r--internal-flume/conf/k2f/pro/kafka.properties25
-rw-r--r--internal-flume/conf/k2f/pro/kafka1.properties25
-rw-r--r--internal-flume/conf/k2f/pro/kafka2.properties25
-rw-r--r--internal-flume/conf/k2f/pro/log4j.properties68
-rw-r--r--internal-flume/conf/k2f/pro/tail.properties34
-rw-r--r--internal-flume/conf/k2f/restart_log/restart_f2k.log213
-rw-r--r--internal-flume/conf/k2f/restart_log/restart_security_f2k.log134
-rwxr-xr-xinternal-flume/conf/k2f/start_flume.sh5
-rw-r--r--internal-flume/conf/proxy/restart_log/restart_proxy_f2k.log34
-rwxr-xr-xinternal-flume/conf/proxy/start_flume.sh5
-rw-r--r--internal-flume/conf/radius/restart_log/restart_radius_f2k.log28
-rwxr-xr-xinternal-flume/conf/radius/start_flume.sh5
-rwxr-xr-xinternal-flume/conf/security/start_flume.sh5
-rw-r--r--internal-flume/origin/conf/active/active_defence_f2k1.properties40
-rw-r--r--internal-flume/origin/conf/active/active_defence_f2k2.properties40
-rw-r--r--internal-flume/origin/conf/active/active_defence_f2k3.properties40
-rw-r--r--internal-flume/origin/conf/active/active_defence_f2k4.properties40
-rw-r--r--internal-flume/origin/conf/active/active_defence_f2k5.properties40
-rwxr-xr-xinternal-flume/origin/conf/active/clean_start_flume.sh (renamed from internal-flume/conf/k2f/clean_start_flume.sh)0
-rwxr-xr-xinternal-flume/origin/conf/active/count_flume.sh (renamed from internal-flume/conf/proxy/count_flume.sh)0
-rwxr-xr-xinternal-flume/origin/conf/active/dae_active_denfence.sh39
-rwxr-xr-xinternal-flume/origin/conf/active/flume-env.sh (renamed from internal-flume/conf/k2f/flume-env.sh)0
-rw-r--r--internal-flume/origin/conf/active/restart_log/restart_connection_f2k.log96
-rwxr-xr-xinternal-flume/origin/conf/active/start_flume.sh4
-rwxr-xr-xinternal-flume/origin/conf/active/stop_flume.sh (renamed from internal-flume/conf/proxy/stop_flume.sh)0
-rwxr-xr-xinternal-flume/origin/conf/connection/clean_start_flume.sh (renamed from internal-flume/conf/proxy/clean_start_flume.sh)0
-rw-r--r--internal-flume/origin/conf/connection/connection_f2k1.properties (renamed from internal-flume/conf/connection/connection_f2k1.properties)2
-rw-r--r--internal-flume/origin/conf/connection/connection_f2k2.properties (renamed from internal-flume/conf/connection/connection_f2k2.properties)2
-rw-r--r--internal-flume/origin/conf/connection/connection_f2k3.properties (renamed from internal-flume/conf/connection/connection_f2k3.properties)2
-rw-r--r--internal-flume/origin/conf/connection/connection_f2k4.properties (renamed from internal-flume/conf/connection/connection_f2k4.properties)2
-rw-r--r--internal-flume/origin/conf/connection/connection_f2k5.properties (renamed from internal-flume/conf/connection/connection_f2k5.properties)2
-rwxr-xr-xinternal-flume/origin/conf/connection/count_flume.sh (renamed from internal-flume/conf/radius/count_flume.sh)0
-rwxr-xr-xinternal-flume/origin/conf/connection/dae_connection.sh (renamed from internal-flume/conf/connection/dae_connection.sh)0
-rwxr-xr-xinternal-flume/origin/conf/connection/flume-env.sh (renamed from internal-flume/conf/proxy/flume-env.sh)0
-rw-r--r--internal-flume/origin/conf/connection/restart_log/restart_connection_f2k.log96
-rwxr-xr-xinternal-flume/origin/conf/connection/start_flume.sh (renamed from internal-flume/conf/connection/start_flume.sh)0
-rwxr-xr-xinternal-flume/origin/conf/connection/stop_flume.sh (renamed from internal-flume/conf/radius/stop_flume.sh)0
-rwxr-xr-xinternal-flume/origin/conf/proxy/clean_start_flume.sh (renamed from internal-flume/conf/radius/clean_start_flume.sh)0
-rwxr-xr-xinternal-flume/origin/conf/proxy/count_flume.sh (renamed from internal-flume/conf/security/count_flume.sh)0
-rwxr-xr-xinternal-flume/origin/conf/proxy/dae_proxy.sh (renamed from internal-flume/conf/proxy/dae_proxy.sh)0
-rwxr-xr-xinternal-flume/origin/conf/proxy/flume-env.sh (renamed from internal-flume/conf/radius/flume-env.sh)0
-rw-r--r--internal-flume/origin/conf/proxy/proxy_f2k1.properties (renamed from internal-flume/conf/proxy/proxy_f2k1.properties)2
-rw-r--r--internal-flume/origin/conf/proxy/proxy_f2k2.properties (renamed from internal-flume/conf/proxy/proxy_f2k2.properties)2
-rw-r--r--internal-flume/origin/conf/proxy/proxy_f2k3.properties (renamed from internal-flume/conf/proxy/proxy_f2k3.properties)2
-rw-r--r--internal-flume/origin/conf/proxy/proxy_f2k4.properties (renamed from internal-flume/conf/proxy/proxy_f2k4.properties)2
-rw-r--r--internal-flume/origin/conf/proxy/proxy_f2k5.properties (renamed from internal-flume/conf/proxy/proxy_f2k5.properties)2
-rw-r--r--internal-flume/origin/conf/proxy/restart_log/restart_proxy_f2k.log74
-rwxr-xr-xinternal-flume/origin/conf/proxy/start_flume.sh5
-rwxr-xr-xinternal-flume/origin/conf/proxy/stop_flume.sh (renamed from internal-flume/conf/security/stop_flume.sh)0
-rwxr-xr-xinternal-flume/origin/conf/radius/clean_start_flume.sh (renamed from internal-flume/conf/security/clean_start_flume.sh)0
-rwxr-xr-xinternal-flume/origin/conf/radius/count_flume.sh (renamed from internal-flume/conf/k2f/count_flume.sh)4
-rwxr-xr-xinternal-flume/origin/conf/radius/dae_radius.sh (renamed from internal-flume/conf/radius/dae_radius.sh)0
-rwxr-xr-xinternal-flume/origin/conf/radius/flume-env.sh (renamed from internal-flume/conf/security/flume-env.sh)0
-rw-r--r--internal-flume/origin/conf/radius/radius_f2k1.properties (renamed from internal-flume/conf/radius/radius_f2k1.properties)2
-rw-r--r--internal-flume/origin/conf/radius/radius_f2k2.properties (renamed from internal-flume/conf/radius/radius_f2k2.properties)2
-rw-r--r--internal-flume/origin/conf/radius/radius_f2k3.properties (renamed from internal-flume/conf/radius/radius_f2k3.properties)2
-rw-r--r--internal-flume/origin/conf/radius/radius_f2k4.properties (renamed from internal-flume/conf/radius/radius_f2k4.properties)2
-rw-r--r--internal-flume/origin/conf/radius/radius_f2k5.properties (renamed from internal-flume/conf/radius/radius_f2k5.properties)2
-rw-r--r--internal-flume/origin/conf/radius/restart_log/restart_radius_f2k.log68
-rwxr-xr-xinternal-flume/origin/conf/radius/start_flume.sh5
-rwxr-xr-xinternal-flume/origin/conf/radius/stop_flume.sh (renamed from internal-flume/conf/k2f/stop_flume.sh)9
-rwxr-xr-xinternal-flume/origin/conf/security/clean_start_flume.sh10
-rwxr-xr-xinternal-flume/origin/conf/security/count_flume.sh15
-rwxr-xr-xinternal-flume/origin/conf/security/dae_security.sh (renamed from internal-flume/conf/security/dae_security.sh)0
-rwxr-xr-xinternal-flume/origin/conf/security/flume-env.sh35
-rw-r--r--internal-flume/origin/conf/security/restart_log/restart_security_f2k.log (renamed from internal-flume/conf/security/restart_log/restart_security_f2k.log)45
-rw-r--r--internal-flume/origin/conf/security/security_f2k1.properties (renamed from internal-flume/conf/security/security_f2k1.properties)2
-rw-r--r--internal-flume/origin/conf/security/security_f2k2.properties (renamed from internal-flume/conf/security/security_f2k2.properties)2
-rw-r--r--internal-flume/origin/conf/security/security_f2k3.properties (renamed from internal-flume/conf/security/security_f2k3.properties)2
-rw-r--r--internal-flume/origin/conf/security/security_f2k4.properties (renamed from internal-flume/conf/security/security_f2k4.properties)2
-rw-r--r--internal-flume/origin/conf/security/security_f2k5.properties (renamed from internal-flume/conf/security/security_f2k5.properties)2
-rwxr-xr-xinternal-flume/origin/conf/security/start_flume.sh5
-rwxr-xr-xinternal-flume/origin/conf/security/stop_flume.sh31
-rw-r--r--internal-flume/origin/docker-compose.yml (renamed from internal-flume/docker-compose.yml)0
-rw-r--r--internal-flume/origin/flumeController/bootstart.log (renamed from internal-flume/flumeController/bootstart.log)0
-rwxr-xr-xinternal-flume/origin/flumeController/change_log_lv.sh (renamed from internal-flume/flumeController/change_log_lv.sh)0
-rwxr-xr-xinternal-flume/origin/flumeController/clean_log_all.sh (renamed from internal-flume/flumeController/clean_log_all.sh)0
-rwxr-xr-xinternal-flume/origin/flumeController/clean_start_all.sh (renamed from internal-flume/flumeController/clean_start_all.sh)0
-rwxr-xr-xinternal-flume/origin/flumeController/count_all.sh (renamed from internal-flume/flumeController/count_all.sh)0
-rwxr-xr-xinternal-flume/origin/flumeController/start_all.sh (renamed from internal-flume/flumeController/start_all.sh)0
-rwxr-xr-xinternal-flume/origin/flumeController/stop_all.sh (renamed from internal-flume/flumeController/stop_all.sh)0
-rw-r--r--internal-flume/origin/flumeMonitor/file-monitor-flume.log (renamed from internal-flume/flumeMonitor/file-monitor-flume.log)0
-rw-r--r--internal-flume/origin/flumeMonitor/flumeConfList (renamed from internal-flume/flumeMonitor/flumeConfList)0
-rw-r--r--internal-flume/origin/flumeMonitor/readMe.txt (renamed from internal-flume/flumeMonitor/readMe.txt)0
-rwxr-xr-xinternal-flume/origin/flumeMonitor/taskFlumeMinSum.sh (renamed from internal-flume/flumeMonitor/taskFlumeMinSum.sh)0
227 files changed, 2450 insertions, 964 deletions
diff --git a/external-flume/conf/radius/radius_k2f.properties b/external-flume/conf/radius/radius_k2f.properties
deleted file mode 100644
index 895414f..0000000
--- a/external-flume/conf/radius/radius_k2f.properties
+++ /dev/null
@@ -1,22 +0,0 @@
-#为source channel sink起名
-radius.channels = c1
-radius.sinks = k1
-#指定channel为kafka channel,省略source使得效率更高
-radius.channels.c1.type = org.apache.flume.channel.kafka.KafkaChannel
-radius.channels.c1.kafka.bootstrap.servers = 192.168.40.152:9092
-radius.channels.c1.kafka.topic = RADIUS-RECORD-LOG
-radius.channels.c1.kafka.consumer.group.id = lxk_0514
-radius.channels.c1.kafka.consumer.auto.offset.reset = latest
-radius.channels.c1.kafka.consumer.enable.auto.commit = true
-radius.channels.c1.kafka.consumer.fetch.max.wait.ms = 1000
-radius.channels.c1.kafka.consumer.fetch.min.bytes = 1048576
-radius.channels.c1.parseAsFlumeEvent = false
-
-#原始file roll sink,将数据写到本机磁盘
-radius.sinks.k1.type = file_roll
-radius.sinks.k1.channel = c1
-radius.sinks.k1.sink.pathManager = default
-radius.sinks.k1.sink.pathManager.extension = dat
-radius.sinks.k1.sink.pathManager.prefix = test-
-radius.sinks.k1.sink.rollInterval = 30
-radius.sinks.k1.sink.directory = /home/192.168.60.101/RADIUS-RECORD-LOG
diff --git a/external-flume/conf_external/active/active_defence_k2f.properties b/external-flume/conf_external/active/active_defence_k2f.properties
new file mode 100644
index 0000000..77a2ae9
--- /dev/null
+++ b/external-flume/conf_external/active/active_defence_k2f.properties
@@ -0,0 +1,22 @@
+#为source channel sink起名
+active_denfence.channels = c1
+active_denfence.sinks = k1
+#指定channel为kafka channel,省略source使得效率更高
+active_denfence.channels.c1.type = org.apache.flume.channel.kafka.KafkaChannel
+active_denfence.channels.c1.kafka.bootstrap.servers = 192.168.40.152:9092
+active_denfence.channels.c1.kafka.topic = ACTIVE-DEFENCE-EVENT-LOG
+active_denfence.channels.c1.kafka.consumer.group.id = lxk_0514
+active_denfence.channels.c1.kafka.consumer.auto.offset.reset = latest
+active_denfence.channels.c1.kafka.consumer.enable.auto.commit = true
+active_denfence.channels.c1.kafka.consumer.fetch.max.wait.ms = 100
+active_denfence.channels.c1.kafka.consumer.fetch.min.bytes = 1048576
+active_denfence.channels.c1.parseAsFlumeEvent = false
+
+#原始file roll sink,将数据写到本机磁盘
+active_denfence.sinks.k1.type = file_roll
+active_denfence.sinks.k1.channel = c1
+active_denfence.sinks.k1.sink.pathManager = default
+active_denfence.sinks.k1.sink.pathManager.extension = dat
+active_denfence.sinks.k1.sink.pathManager.prefix = test-
+active_denfence.sinks.k1.sink.rollInterval = 30
+active_denfence.sinks.k1.sink.directory = /home/192.168.60.101/ACTIVE-DEFENCE-EVENT-LOG
diff --git a/external-flume/conf/connection/clean_start_flume.sh b/external-flume/conf_external/active/clean_start_flume.sh
index ef811a5..ef811a5 100755
--- a/external-flume/conf/connection/clean_start_flume.sh
+++ b/external-flume/conf_external/active/clean_start_flume.sh
diff --git a/internal-flume/conf/connection/count_flume.sh b/external-flume/conf_external/active/count_flume.sh
index 4a50a2a..4a50a2a 100755
--- a/internal-flume/conf/connection/count_flume.sh
+++ b/external-flume/conf_external/active/count_flume.sh
diff --git a/internal-flume/conf/k2f/dae_f2k.sh b/external-flume/conf_external/active/dae_active_denfence.sh
index 0c61180..5c9eebd 100755
--- a/internal-flume/conf/k2f/dae_f2k.sh
+++ b/external-flume/conf_external/active/dae_active_denfence.sh
@@ -1,15 +1,15 @@
#!/bin/sh
#JAR_NAME=ktk_ip_asn.properties
-PROPERTIES_NAME=security_f2k
+PROPERTIES_NAME=active_denfence_k2f
#flume进程名称
-FLUME_NAME=security
+FLUME_NAME=active_denfence
#flume根目录
BASE_DIR=$(cd $(dirname $0); cd ../../; pwd)
#当前路径
CONF_DIR=$(cd $(dirname $0); pwd)
#进程总数
-PROCESS_SUM=3
+PROCESS_SUM=5
echo "##############################################################" >> ${CONF_DIR}/restart_log/restart_${PROPERTIES_NAME}.log
id=0 #配置文件后面的序号,无特殊作用
diff --git a/external-flume/conf/connection/flume-env.sh b/external-flume/conf_external/active/flume-env.sh
index 63979aa..63979aa 100755
--- a/external-flume/conf/connection/flume-env.sh
+++ b/external-flume/conf_external/active/flume-env.sh
diff --git a/internal-flume/conf/connection/restart_log/restart_connection_f2k.log b/external-flume/conf_external/active/restart_log/restart_connection_f2k.log
index f5ec36f..ce2b37d 100644
--- a/internal-flume/conf/connection/restart_log/restart_connection_f2k.log
+++ b/external-flume/conf_external/active/restart_log/restart_connection_f2k.log
@@ -55,3 +55,42 @@
20200513102434 ---> the connection_f2k_APP restart ---> 3
20200513102444 ---> the connection_f2k_APP restart ---> 4
20200513102454 ---> the connection_f2k_APP restart ---> 5
+20200513104430 ---> the connection_f2k_APP restart ---> 1
+20200513104440 ---> the connection_f2k_APP restart ---> 2
+20200513104450 ---> the connection_f2k_APP restart ---> 3
+20200513104500 ---> the connection_f2k_APP restart ---> 4
+20200513104510 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200513181554 ---> the connection_f2k_APP restart ---> 1
+20200513181604 ---> the connection_f2k_APP restart ---> 2
+20200513181614 ---> the connection_f2k_APP restart ---> 3
+20200513181624 ---> the connection_f2k_APP restart ---> 4
+20200513181634 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200513182119 ---> the connection_f2k_APP restart ---> 1
+20200513182129 ---> the connection_f2k_APP restart ---> 2
+20200513182139 ---> the connection_f2k_APP restart ---> 3
+20200513182149 ---> the connection_f2k_APP restart ---> 4
+20200513182159 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200513182640 ---> the connection_f2k_APP restart ---> 1
+20200513182651 ---> the connection_f2k_APP restart ---> 2
+20200513182701 ---> the connection_f2k_APP restart ---> 3
+20200513182711 ---> the connection_f2k_APP restart ---> 4
+20200513182721 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200514171459 ---> the connection_f2k_APP restart ---> 1
+20200514171509 ---> the connection_f2k_APP restart ---> 2
+20200514171519 ---> the connection_f2k_APP restart ---> 3
+20200514171529 ---> the connection_f2k_APP restart ---> 4
+20200514171539 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200514172152 ---> the connection_f2k_APP restart ---> 1
+20200514172202 ---> the connection_f2k_APP restart ---> 2
+20200514172212 ---> the connection_f2k_APP restart ---> 3
+##############################################################
+20200514172234 ---> the connection_f2k_APP restart ---> 1
+20200514172244 ---> the connection_f2k_APP restart ---> 2
+20200514172254 ---> the connection_f2k_APP restart ---> 3
+20200514172304 ---> the connection_f2k_APP restart ---> 4
+20200514172314 ---> the connection_f2k_APP restart ---> 5
diff --git a/external-flume/conf/connection/start_flume.sh b/external-flume/conf_external/active/start_flume.sh
index 01dde8b..01dde8b 100755
--- a/external-flume/conf/connection/start_flume.sh
+++ b/external-flume/conf_external/active/start_flume.sh
diff --git a/internal-flume/conf/connection/stop_flume.sh b/external-flume/conf_external/active/stop_flume.sh
index 1d642b6..1d642b6 100755
--- a/internal-flume/conf/connection/stop_flume.sh
+++ b/external-flume/conf_external/active/stop_flume.sh
diff --git a/external-flume/conf/proxy/clean_start_flume.sh b/external-flume/conf_external/connection/clean_start_flume.sh
index ef811a5..ef811a5 100755
--- a/external-flume/conf/proxy/clean_start_flume.sh
+++ b/external-flume/conf_external/connection/clean_start_flume.sh
diff --git a/external-flume/conf/connection/connection_k2f.properties b/external-flume/conf_external/connection/connection_k2f.properties
index f6ca65d..f6ca65d 100644
--- a/external-flume/conf/connection/connection_k2f.properties
+++ b/external-flume/conf_external/connection/connection_k2f.properties
diff --git a/external-flume/conf/connection/count_flume.sh b/external-flume/conf_external/connection/count_flume.sh
index cc8452b..cc8452b 100755
--- a/external-flume/conf/connection/count_flume.sh
+++ b/external-flume/conf_external/connection/count_flume.sh
diff --git a/external-flume/conf/connection/dae_connection.sh b/external-flume/conf_external/connection/dae_connection.sh
index b7ba0d9..b7ba0d9 100755
--- a/external-flume/conf/connection/dae_connection.sh
+++ b/external-flume/conf_external/connection/dae_connection.sh
diff --git a/external-flume/conf/proxy/flume-env.sh b/external-flume/conf_external/connection/flume-env.sh
index 63979aa..63979aa 100755
--- a/external-flume/conf/proxy/flume-env.sh
+++ b/external-flume/conf_external/connection/flume-env.sh
diff --git a/external-flume/conf/connection/log4j.properties b/external-flume/conf_external/connection/log4j.properties
index c948bff..c948bff 100644
--- a/external-flume/conf/connection/log4j.properties
+++ b/external-flume/conf_external/connection/log4j.properties
diff --git a/external-flume/conf_external/connection/logs/flume.log b/external-flume/conf_external/connection/logs/flume.log
new file mode 100644
index 0000000..cb74ece
--- /dev/null
+++ b/external-flume/conf_external/connection/logs/flume.log
@@ -0,0 +1,201 @@
+14 May 2020 11:46:42,409 INFO [lifecycleSupervisor-1-0] (org.apache.flume.node.PollingPropertiesFileConfigurationProvider.start:62) - Configuration provider starting
+14 May 2020 11:46:42,421 INFO [conf-file-poller-0] (org.apache.flume.node.PollingPropertiesFileConfigurationProvider$FileWatcherRunnable.run:138) - Reloading configuration file:/home/apache-flume-1.9.0-bin/conf/connection/connection_k2f.properties
+14 May 2020 11:46:42,431 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,433 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,433 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,434 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,434 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,435 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,435 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1117) - Added sinks: k1 Agent: connection
+14 May 2020 11:46:42,435 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,436 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,436 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,436 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,437 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,437 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,438 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,440 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,440 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,441 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,441 WARN [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateConfigFilterSet:623) - Agent configuration for 'connection' has no configfilters.
+14 May 2020 11:46:42,464 WARN [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateSources:702) - Agent configuration for 'connection' has no sources.
+14 May 2020 11:46:42,469 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration.validateConfiguration:163) - Post-validation flume configuration contains configuration for agents: [connection]
+14 May 2020 11:46:42,469 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:151) - Creating channels
+14 May 2020 11:46:42,481 INFO [conf-file-poller-0] (org.apache.flume.channel.DefaultChannelFactory.create:42) - Creating instance of channel c1 type org.apache.flume.channel.kafka.KafkaChannel
+14 May 2020 11:46:42,503 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:205) - Created channel c1
+14 May 2020 11:46:42,504 INFO [conf-file-poller-0] (org.apache.flume.sink.DefaultSinkFactory.create:42) - Creating instance of sink: k1, type: file_roll
+14 May 2020 11:46:42,514 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.getConfiguration:120) - Channel c1 connected to [k1]
+14 May 2020 11:46:42,524 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:162) - Starting new configuration:{ sourceRunners:{} sinkRunners:{k1=SinkRunner: { policy:org.apache.flume.sink.DefaultSinkProcessor@5cf2aeb7 counterGroup:{ name:null counters:{} } }} channels:{c1=org.apache.flume.channel.kafka.KafkaChannel{name: c1}} }
+14 May 2020 11:46:42,525 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:169) - Starting Channel c1
+14 May 2020 11:46:42,526 INFO [lifecycleSupervisor-1-0] (org.apache.flume.channel.kafka.KafkaChannel.start:135) - Starting Kafka Channel: c1
+14 May 2020 11:46:42,562 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.config.AbstractConfig.logAll:279) - ProducerConfig values:
+ acks = all
+ batch.size = 16384
+ bootstrap.servers = [192.168.40.152:9092]
+ buffer.memory = 33554432
+ client.id =
+ compression.type = none
+ connections.max.idle.ms = 540000
+ enable.idempotence = false
+ interceptor.classes = []
+ key.serializer = class org.apache.kafka.common.serialization.StringSerializer
+ linger.ms = 0
+ max.block.ms = 60000
+ max.in.flight.requests.per.connection = 5
+ max.request.size = 1048576
+ metadata.max.age.ms = 300000
+ metric.reporters = []
+ metrics.num.samples = 2
+ metrics.recording.level = INFO
+ metrics.sample.window.ms = 30000
+ partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
+ receive.buffer.bytes = 32768
+ reconnect.backoff.max.ms = 1000
+ reconnect.backoff.ms = 50
+ request.timeout.ms = 30000
+ retries = 0
+ retry.backoff.ms = 100
+ sasl.client.callback.handler.class = null
+ sasl.jaas.config = null
+ sasl.kerberos.kinit.cmd = /usr/bin/kinit
+ sasl.kerberos.min.time.before.relogin = 60000
+ sasl.kerberos.service.name = null
+ sasl.kerberos.ticket.renew.jitter = 0.05
+ sasl.kerberos.ticket.renew.window.factor = 0.8
+ sasl.login.callback.handler.class = null
+ sasl.login.class = null
+ sasl.login.refresh.buffer.seconds = 300
+ sasl.login.refresh.min.period.seconds = 60
+ sasl.login.refresh.window.factor = 0.8
+ sasl.login.refresh.window.jitter = 0.05
+ sasl.mechanism = GSSAPI
+ security.protocol = PLAINTEXT
+ send.buffer.bytes = 131072
+ ssl.cipher.suites = null
+ ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
+ ssl.endpoint.identification.algorithm = https
+ ssl.key.password = null
+ ssl.keymanager.algorithm = SunX509
+ ssl.keystore.location = null
+ ssl.keystore.password = null
+ ssl.keystore.type = JKS
+ ssl.protocol = TLS
+ ssl.provider = null
+ ssl.secure.random.implementation = null
+ ssl.trustmanager.algorithm = PKIX
+ ssl.truststore.location = null
+ ssl.truststore.password = null
+ ssl.truststore.type = JKS
+ transaction.timeout.ms = 60000
+ transactional.id = null
+ value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
+
+14 May 2020 11:46:42,638 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:109) - Kafka version : 2.0.1
+14 May 2020 11:46:42,639 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:110) - Kafka commitId : fa14705e51bd2ce5
+14 May 2020 11:46:42,641 INFO [lifecycleSupervisor-1-0] (org.apache.flume.channel.kafka.KafkaChannel.start:143) - Topic = CONNECTION-RECORD-LOG
+14 May 2020 11:46:42,641 INFO [lifecycleSupervisor-1-0] (org.apache.flume.instrumentation.MonitoredCounterGroup.register:119) - Monitored counter group for type: CHANNEL, name: c1: Successfully registered new MBean.
+14 May 2020 11:46:42,642 INFO [lifecycleSupervisor-1-0] (org.apache.flume.instrumentation.MonitoredCounterGroup.start:95) - Component type: CHANNEL, name: c1 started
+14 May 2020 11:46:42,642 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:196) - Starting Sink k1
+14 May 2020 11:46:42,643 INFO [lifecycleSupervisor-1-1] (org.apache.flume.sink.RollingFileSink.start:111) - Starting org.apache.flume.sink.RollingFileSink{name:k1, channel:c1}...
+14 May 2020 11:46:42,644 INFO [lifecycleSupervisor-1-1] (org.apache.flume.instrumentation.MonitoredCounterGroup.register:119) - Monitored counter group for type: SINK, name: k1: Successfully registered new MBean.
+14 May 2020 11:46:42,644 INFO [lifecycleSupervisor-1-1] (org.apache.flume.instrumentation.MonitoredCounterGroup.start:95) - Component type: SINK, name: k1 started
+14 May 2020 11:46:42,645 INFO [lifecycleSupervisor-1-1] (org.apache.flume.sink.RollingFileSink.start:143) - RollingFileSink k1 started.
+14 May 2020 11:46:42,661 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.config.AbstractConfig.logAll:279) - ConsumerConfig values:
+ auto.commit.interval.ms = 5000
+ auto.offset.reset = latest
+ bootstrap.servers = [192.168.40.152:9092]
+ check.crcs = true
+ client.id =
+ connections.max.idle.ms = 540000
+ default.api.timeout.ms = 60000
+ enable.auto.commit = false
+ exclude.internal.topics = true
+ fetch.max.bytes = 52428800
+ fetch.max.wait.ms = 1000
+ fetch.min.bytes = 1048576
+ group.id = lxk_0514
+ heartbeat.interval.ms = 3000
+ interceptor.classes = []
+ internal.leave.group.on.close = true
+ isolation.level = read_uncommitted
+ key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
+ max.partition.fetch.bytes = 1048576
+ max.poll.interval.ms = 300000
+ max.poll.records = 500
+ metadata.max.age.ms = 300000
+ metric.reporters = []
+ metrics.num.samples = 2
+ metrics.recording.level = INFO
+ metrics.sample.window.ms = 30000
+ partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
+ receive.buffer.bytes = 65536
+ reconnect.backoff.max.ms = 1000
+ reconnect.backoff.ms = 50
+ request.timeout.ms = 30000
+ retry.backoff.ms = 100
+ sasl.client.callback.handler.class = null
+ sasl.jaas.config = null
+ sasl.kerberos.kinit.cmd = /usr/bin/kinit
+ sasl.kerberos.min.time.before.relogin = 60000
+ sasl.kerberos.service.name = null
+ sasl.kerberos.ticket.renew.jitter = 0.05
+ sasl.kerberos.ticket.renew.window.factor = 0.8
+ sasl.login.callback.handler.class = null
+ sasl.login.class = null
+ sasl.login.refresh.buffer.seconds = 300
+ sasl.login.refresh.min.period.seconds = 60
+ sasl.login.refresh.window.factor = 0.8
+ sasl.login.refresh.window.jitter = 0.05
+ sasl.mechanism = GSSAPI
+ security.protocol = PLAINTEXT
+ send.buffer.bytes = 131072
+ session.timeout.ms = 10000
+ ssl.cipher.suites = null
+ ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
+ ssl.endpoint.identification.algorithm = https
+ ssl.key.password = null
+ ssl.keymanager.algorithm = SunX509
+ ssl.keystore.location = null
+ ssl.keystore.password = null
+ ssl.keystore.type = JKS
+ ssl.protocol = TLS
+ ssl.provider = null
+ ssl.secure.random.implementation = null
+ ssl.trustmanager.algorithm = PKIX
+ ssl.truststore.location = null
+ ssl.truststore.password = null
+ ssl.truststore.type = JKS
+ value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
+
+14 May 2020 11:46:42,696 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:109) - Kafka version : 2.0.1
+14 May 2020 11:46:42,697 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:110) - Kafka commitId : fa14705e51bd2ce5
+14 May 2020 11:46:42,700 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.KafkaChannel.createConsumerAndRecords:304) - Created new consumer to connect to Kafka
+14 May 2020 11:46:42,870 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.Metadata.update:285) - Cluster ID: gsZT8xNUSViLGeSpG1neHg
+14 May 2020 11:46:42,871 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$FindCoordinatorResponseHandler.onSuccess:677) - [Consumer clientId=consumer-1, groupId=lxk_0514] Discovered group coordinator 192.168.40.152:9092 (id: 2147483646 rack: null)
+14 May 2020 11:46:42,873 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions []
+14 May 2020 11:46:42,873 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 11:46:42,887 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 18
+14 May 2020 11:46:42,888 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 11:46:42,889 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic CONNECTION-RECORD-LOG - partition 0 assigned.
+14 May 2020 11:46:45,901 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$HeartbeatResponseHandler.handle:855) - [Consumer clientId=consumer-1, groupId=lxk_0514] Attempt to heartbeat failed since group is rebalancing
+14 May 2020 11:46:45,910 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 11:46:45,910 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsRevoked:785) - topic CONNECTION-RECORD-LOG - partition 0 revoked.
+14 May 2020 11:46:45,911 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 11:46:45,921 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 19
+14 May 2020 11:46:45,922 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 11:46:45,923 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic CONNECTION-RECORD-LOG - partition 0 assigned.
+14 May 2020 11:51:42,641 INFO [kafka-producer-network-thread | producer-1] (org.apache.kafka.clients.Metadata.update:285) - Cluster ID: gsZT8xNUSViLGeSpG1neHg
+14 May 2020 12:02:48,119 INFO [kafka-coordinator-heartbeat-thread | lxk_0514] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$HeartbeatResponseHandler.handle:855) - [Consumer clientId=consumer-1, groupId=lxk_0514] Attempt to heartbeat failed since group is rebalancing
+14 May 2020 12:02:48,715 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 12:02:48,716 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsRevoked:785) - topic CONNECTION-RECORD-LOG - partition 0 revoked.
+14 May 2020 12:02:48,717 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 12:02:54,224 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 20
+14 May 2020 12:02:54,225 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 12:02:54,225 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic CONNECTION-RECORD-LOG - partition 0 assigned.
+14 May 2020 12:02:57,326 INFO [kafka-coordinator-heartbeat-thread | lxk_0514] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$HeartbeatResponseHandler.handle:855) - [Consumer clientId=consumer-1, groupId=lxk_0514] Attempt to heartbeat failed since group is rebalancing
+14 May 2020 12:02:59,720 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 12:02:59,721 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsRevoked:785) - topic CONNECTION-RECORD-LOG - partition 0 revoked.
+14 May 2020 12:02:59,722 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 12:02:59,729 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 21
+14 May 2020 12:02:59,730 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 12:02:59,730 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic CONNECTION-RECORD-LOG - partition 0 assigned.
diff --git a/external-flume/conf/connection/pro/2f2k.properties b/external-flume/conf_external/connection/pro/2f2k.properties
index 5e5c573..5e5c573 100644
--- a/external-flume/conf/connection/pro/2f2k.properties
+++ b/external-flume/conf_external/connection/pro/2f2k.properties
diff --git a/external-flume/conf/connection/pro/4f2k.properties b/external-flume/conf_external/connection/pro/4f2k.properties
index bf0408b..bf0408b 100644
--- a/external-flume/conf/connection/pro/4f2k.properties
+++ b/external-flume/conf_external/connection/pro/4f2k.properties
diff --git a/external-flume/conf/connection/pro/f2k.properties b/external-flume/conf_external/connection/pro/f2k.properties
index 7047e42..7047e42 100644
--- a/external-flume/conf/connection/pro/f2k.properties
+++ b/external-flume/conf_external/connection/pro/f2k.properties
diff --git a/external-flume/conf/connection/pro/f2k1.properties b/external-flume/conf_external/connection/pro/f2k1.properties
index 2dadced..2dadced 100644
--- a/external-flume/conf/connection/pro/f2k1.properties
+++ b/external-flume/conf_external/connection/pro/f2k1.properties
diff --git a/external-flume/conf/connection/pro/f2k2.properties b/external-flume/conf_external/connection/pro/f2k2.properties
index e17c3e2..e17c3e2 100644
--- a/external-flume/conf/connection/pro/f2k2.properties
+++ b/external-flume/conf_external/connection/pro/f2k2.properties
diff --git a/external-flume/conf/connection/pro/f2k3.properties b/external-flume/conf_external/connection/pro/f2k3.properties
index 2dadced..2dadced 100644
--- a/external-flume/conf/connection/pro/f2k3.properties
+++ b/external-flume/conf_external/connection/pro/f2k3.properties
diff --git a/external-flume/conf/connection/pro/file.properties b/external-flume/conf_external/connection/pro/file.properties
index f6813e7..f6813e7 100644
--- a/external-flume/conf/connection/pro/file.properties
+++ b/external-flume/conf_external/connection/pro/file.properties
diff --git a/external-flume/conf/connection/pro/file1.properties b/external-flume/conf_external/connection/pro/file1.properties
index 17dac7f..17dac7f 100644
--- a/external-flume/conf/connection/pro/file1.properties
+++ b/external-flume/conf_external/connection/pro/file1.properties
diff --git a/external-flume/conf/connection/pro/file2.properties b/external-flume/conf_external/connection/pro/file2.properties
index f6813e7..f6813e7 100644
--- a/external-flume/conf/connection/pro/file2.properties
+++ b/external-flume/conf_external/connection/pro/file2.properties
diff --git a/external-flume/conf/connection/pro/k2f.properties b/external-flume/conf_external/connection/pro/k2f.properties
index 35d7d68..35d7d68 100644
--- a/external-flume/conf/connection/pro/k2f.properties
+++ b/external-flume/conf_external/connection/pro/k2f.properties
diff --git a/external-flume/conf/connection/pro/kafka.properties b/external-flume/conf_external/connection/pro/kafka.properties
index da560b2..da560b2 100644
--- a/external-flume/conf/connection/pro/kafka.properties
+++ b/external-flume/conf_external/connection/pro/kafka.properties
diff --git a/external-flume/conf/connection/pro/kafka1.properties b/external-flume/conf_external/connection/pro/kafka1.properties
index ed30b16..ed30b16 100644
--- a/external-flume/conf/connection/pro/kafka1.properties
+++ b/external-flume/conf_external/connection/pro/kafka1.properties
diff --git a/external-flume/conf/connection/pro/kafka2.properties b/external-flume/conf_external/connection/pro/kafka2.properties
index 6040a05..6040a05 100644
--- a/external-flume/conf/connection/pro/kafka2.properties
+++ b/external-flume/conf_external/connection/pro/kafka2.properties
diff --git a/external-flume/conf/connection/pro/log4j.properties b/external-flume/conf_external/connection/pro/log4j.properties
index c948bff..c948bff 100644
--- a/external-flume/conf/connection/pro/log4j.properties
+++ b/external-flume/conf_external/connection/pro/log4j.properties
diff --git a/external-flume/conf/connection/pro/tail.properties b/external-flume/conf_external/connection/pro/tail.properties
index c677318..c677318 100644
--- a/external-flume/conf/connection/pro/tail.properties
+++ b/external-flume/conf_external/connection/pro/tail.properties
diff --git a/external-flume/conf/radius/restart_log/restart_connection_k2f.log b/external-flume/conf_external/connection/restart_log/restart_connection_k2f.log
index 281d596..1ad7998 100644
--- a/external-flume/conf/radius/restart_log/restart_connection_k2f.log
+++ b/external-flume/conf_external/connection/restart_log/restart_connection_k2f.log
@@ -66,3 +66,7 @@
20200513192014 ---> the connection_k2f_APP restart ---> 1
##############################################################
20200513192212 ---> the connection_k2f_APP restart ---> 1
+##############################################################
+20200514112452 ---> the connection_k2f_APP restart ---> 1
+##############################################################
+20200514114641 ---> the connection_k2f_APP restart ---> 1
diff --git a/external-flume/conf/connection/restart_log/restart_f2k.log b/external-flume/conf_external/connection/restart_log/restart_f2k.log
index 9d5fe99..9d5fe99 100644
--- a/external-flume/conf/connection/restart_log/restart_f2k.log
+++ b/external-flume/conf_external/connection/restart_log/restart_f2k.log
diff --git a/external-flume/conf/connection/restart_log/restart_security_f2k.log b/external-flume/conf_external/connection/restart_log/restart_security_f2k.log
index 690385d..690385d 100644
--- a/external-flume/conf/connection/restart_log/restart_security_f2k.log
+++ b/external-flume/conf_external/connection/restart_log/restart_security_f2k.log
diff --git a/external-flume/conf/proxy/start_flume.sh b/external-flume/conf_external/connection/start_flume.sh
index 01dde8b..01dde8b 100755
--- a/external-flume/conf/proxy/start_flume.sh
+++ b/external-flume/conf_external/connection/start_flume.sh
diff --git a/external-flume/conf/connection/stop_flume.sh b/external-flume/conf_external/connection/stop_flume.sh
index e4417aa..e4417aa 100755
--- a/external-flume/conf/connection/stop_flume.sh
+++ b/external-flume/conf_external/connection/stop_flume.sh
diff --git a/external-flume/conf/radius/clean_start_flume.sh b/external-flume/conf_external/proxy/clean_start_flume.sh
index ef811a5..ef811a5 100755
--- a/external-flume/conf/radius/clean_start_flume.sh
+++ b/external-flume/conf_external/proxy/clean_start_flume.sh
diff --git a/external-flume/conf/proxy/count_flume.sh b/external-flume/conf_external/proxy/count_flume.sh
index cc8452b..cc8452b 100755
--- a/external-flume/conf/proxy/count_flume.sh
+++ b/external-flume/conf_external/proxy/count_flume.sh
diff --git a/external-flume/conf/proxy/dae_proxy.sh b/external-flume/conf_external/proxy/dae_proxy.sh
index 5716c9d..5716c9d 100755
--- a/external-flume/conf/proxy/dae_proxy.sh
+++ b/external-flume/conf_external/proxy/dae_proxy.sh
diff --git a/external-flume/conf/radius/flume-env.sh b/external-flume/conf_external/proxy/flume-env.sh
index 63979aa..63979aa 100755
--- a/external-flume/conf/radius/flume-env.sh
+++ b/external-flume/conf_external/proxy/flume-env.sh
diff --git a/external-flume/conf/proxy/log4j.properties b/external-flume/conf_external/proxy/log4j.properties
index c948bff..c948bff 100644
--- a/external-flume/conf/proxy/log4j.properties
+++ b/external-flume/conf_external/proxy/log4j.properties
diff --git a/external-flume/conf_external/proxy/logs/flume.log b/external-flume/conf_external/proxy/logs/flume.log
new file mode 100644
index 0000000..42f0de3
--- /dev/null
+++ b/external-flume/conf_external/proxy/logs/flume.log
@@ -0,0 +1,359 @@
+14 May 2020 11:46:42,712 INFO [lifecycleSupervisor-1-0] (org.apache.flume.node.PollingPropertiesFileConfigurationProvider.start:62) - Configuration provider starting
+14 May 2020 11:46:42,720 INFO [conf-file-poller-0] (org.apache.flume.node.PollingPropertiesFileConfigurationProvider$FileWatcherRunnable.run:138) - Reloading configuration file:/home/apache-flume-1.9.0-bin/conf/proxy/proxy_k2f.properties
+14 May 2020 11:46:42,725 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,725 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,726 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,726 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,726 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,727 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,727 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,727 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,728 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,728 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,728 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1117) - Added sinks: k1 Agent: proxy
+14 May 2020 11:46:42,728 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,728 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,729 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,730 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,730 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,730 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,731 WARN [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateConfigFilterSet:623) - Agent configuration for 'proxy' has no configfilters.
+14 May 2020 11:46:42,741 WARN [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateSources:702) - Agent configuration for 'proxy' has no sources.
+14 May 2020 11:46:42,743 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration.validateConfiguration:163) - Post-validation flume configuration contains configuration for agents: [proxy]
+14 May 2020 11:46:42,743 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:151) - Creating channels
+14 May 2020 11:46:42,749 INFO [conf-file-poller-0] (org.apache.flume.channel.DefaultChannelFactory.create:42) - Creating instance of channel c1 type org.apache.flume.channel.kafka.KafkaChannel
+14 May 2020 11:46:42,761 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:205) - Created channel c1
+14 May 2020 11:46:42,761 INFO [conf-file-poller-0] (org.apache.flume.sink.DefaultSinkFactory.create:42) - Creating instance of sink: k1, type: file_roll
+14 May 2020 11:46:42,767 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.getConfiguration:120) - Channel c1 connected to [k1]
+14 May 2020 11:46:42,774 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:162) - Starting new configuration:{ sourceRunners:{} sinkRunners:{k1=SinkRunner: { policy:org.apache.flume.sink.DefaultSinkProcessor@7e8b5303 counterGroup:{ name:null counters:{} } }} channels:{c1=org.apache.flume.channel.kafka.KafkaChannel{name: c1}} }
+14 May 2020 11:46:42,774 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:169) - Starting Channel c1
+14 May 2020 11:46:42,775 INFO [lifecycleSupervisor-1-0] (org.apache.flume.channel.kafka.KafkaChannel.start:135) - Starting Kafka Channel: c1
+14 May 2020 11:46:42,799 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.config.AbstractConfig.logAll:279) - ProducerConfig values:
+ acks = all
+ batch.size = 16384
+ bootstrap.servers = [192.168.40.152:9092]
+ buffer.memory = 33554432
+ client.id =
+ compression.type = none
+ connections.max.idle.ms = 540000
+ enable.idempotence = false
+ interceptor.classes = []
+ key.serializer = class org.apache.kafka.common.serialization.StringSerializer
+ linger.ms = 0
+ max.block.ms = 60000
+ max.in.flight.requests.per.connection = 5
+ max.request.size = 1048576
+ metadata.max.age.ms = 300000
+ metric.reporters = []
+ metrics.num.samples = 2
+ metrics.recording.level = INFO
+ metrics.sample.window.ms = 30000
+ partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
+ receive.buffer.bytes = 32768
+ reconnect.backoff.max.ms = 1000
+ reconnect.backoff.ms = 50
+ request.timeout.ms = 30000
+ retries = 0
+ retry.backoff.ms = 100
+ sasl.client.callback.handler.class = null
+ sasl.jaas.config = null
+ sasl.kerberos.kinit.cmd = /usr/bin/kinit
+ sasl.kerberos.min.time.before.relogin = 60000
+ sasl.kerberos.service.name = null
+ sasl.kerberos.ticket.renew.jitter = 0.05
+ sasl.kerberos.ticket.renew.window.factor = 0.8
+ sasl.login.callback.handler.class = null
+ sasl.login.class = null
+ sasl.login.refresh.buffer.seconds = 300
+ sasl.login.refresh.min.period.seconds = 60
+ sasl.login.refresh.window.factor = 0.8
+ sasl.login.refresh.window.jitter = 0.05
+ sasl.mechanism = GSSAPI
+ security.protocol = PLAINTEXT
+ send.buffer.bytes = 131072
+ ssl.cipher.suites = null
+ ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
+ ssl.endpoint.identification.algorithm = https
+ ssl.key.password = null
+ ssl.keymanager.algorithm = SunX509
+ ssl.keystore.location = null
+ ssl.keystore.password = null
+ ssl.keystore.type = JKS
+ ssl.protocol = TLS
+ ssl.provider = null
+ ssl.secure.random.implementation = null
+ ssl.trustmanager.algorithm = PKIX
+ ssl.truststore.location = null
+ ssl.truststore.password = null
+ ssl.truststore.type = JKS
+ transaction.timeout.ms = 60000
+ transactional.id = null
+ value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
+
+14 May 2020 11:46:42,854 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:109) - Kafka version : 2.0.1
+14 May 2020 11:46:42,855 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:110) - Kafka commitId : fa14705e51bd2ce5
+14 May 2020 11:46:42,857 INFO [lifecycleSupervisor-1-0] (org.apache.flume.channel.kafka.KafkaChannel.start:143) - Topic = PROXY-EVENT-LOG
+14 May 2020 11:46:42,857 INFO [lifecycleSupervisor-1-0] (org.apache.flume.instrumentation.MonitoredCounterGroup.register:119) - Monitored counter group for type: CHANNEL, name: c1: Successfully registered new MBean.
+14 May 2020 11:46:42,858 INFO [lifecycleSupervisor-1-0] (org.apache.flume.instrumentation.MonitoredCounterGroup.start:95) - Component type: CHANNEL, name: c1 started
+14 May 2020 11:46:42,858 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:196) - Starting Sink k1
+14 May 2020 11:46:42,859 INFO [lifecycleSupervisor-1-3] (org.apache.flume.sink.RollingFileSink.start:111) - Starting org.apache.flume.sink.RollingFileSink{name:k1, channel:c1}...
+14 May 2020 11:46:42,860 INFO [lifecycleSupervisor-1-3] (org.apache.flume.instrumentation.MonitoredCounterGroup.register:119) - Monitored counter group for type: SINK, name: k1: Successfully registered new MBean.
+14 May 2020 11:46:42,860 INFO [lifecycleSupervisor-1-3] (org.apache.flume.instrumentation.MonitoredCounterGroup.start:95) - Component type: SINK, name: k1 started
+14 May 2020 11:46:42,862 INFO [lifecycleSupervisor-1-3] (org.apache.flume.sink.RollingFileSink.start:143) - RollingFileSink k1 started.
+14 May 2020 11:46:42,884 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.config.AbstractConfig.logAll:279) - ConsumerConfig values:
+ auto.commit.interval.ms = 5000
+ auto.offset.reset = latest
+ bootstrap.servers = [192.168.40.152:9092]
+ check.crcs = true
+ client.id =
+ connections.max.idle.ms = 540000
+ default.api.timeout.ms = 60000
+ enable.auto.commit = false
+ exclude.internal.topics = true
+ fetch.max.bytes = 52428800
+ fetch.max.wait.ms = 1000
+ fetch.min.bytes = 1048576
+ group.id = lxk_0514
+ heartbeat.interval.ms = 3000
+ interceptor.classes = []
+ internal.leave.group.on.close = true
+ isolation.level = read_uncommitted
+ key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
+ max.partition.fetch.bytes = 1048576
+ max.poll.interval.ms = 300000
+ max.poll.records = 500
+ metadata.max.age.ms = 300000
+ metric.reporters = []
+ metrics.num.samples = 2
+ metrics.recording.level = INFO
+ metrics.sample.window.ms = 30000
+ partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
+ receive.buffer.bytes = 65536
+ reconnect.backoff.max.ms = 1000
+ reconnect.backoff.ms = 50
+ request.timeout.ms = 30000
+ retry.backoff.ms = 100
+ sasl.client.callback.handler.class = null
+ sasl.jaas.config = null
+ sasl.kerberos.kinit.cmd = /usr/bin/kinit
+ sasl.kerberos.min.time.before.relogin = 60000
+ sasl.kerberos.service.name = null
+ sasl.kerberos.ticket.renew.jitter = 0.05
+ sasl.kerberos.ticket.renew.window.factor = 0.8
+ sasl.login.callback.handler.class = null
+ sasl.login.class = null
+ sasl.login.refresh.buffer.seconds = 300
+ sasl.login.refresh.min.period.seconds = 60
+ sasl.login.refresh.window.factor = 0.8
+ sasl.login.refresh.window.jitter = 0.05
+ sasl.mechanism = GSSAPI
+ security.protocol = PLAINTEXT
+ send.buffer.bytes = 131072
+ session.timeout.ms = 10000
+ ssl.cipher.suites = null
+ ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
+ ssl.endpoint.identification.algorithm = https
+ ssl.key.password = null
+ ssl.keymanager.algorithm = SunX509
+ ssl.keystore.location = null
+ ssl.keystore.password = null
+ ssl.keystore.type = JKS
+ ssl.protocol = TLS
+ ssl.provider = null
+ ssl.secure.random.implementation = null
+ ssl.trustmanager.algorithm = PKIX
+ ssl.truststore.location = null
+ ssl.truststore.password = null
+ ssl.truststore.type = JKS
+ value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
+
+14 May 2020 11:46:42,934 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:109) - Kafka version : 2.0.1
+14 May 2020 11:46:42,934 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:110) - Kafka commitId : fa14705e51bd2ce5
+14 May 2020 11:46:42,939 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.KafkaChannel.createConsumerAndRecords:304) - Created new consumer to connect to Kafka
+14 May 2020 11:46:43,146 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.Metadata.update:285) - Cluster ID: gsZT8xNUSViLGeSpG1neHg
+14 May 2020 11:46:43,147 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$FindCoordinatorResponseHandler.onSuccess:677) - [Consumer clientId=consumer-1, groupId=lxk_0514] Discovered group coordinator 192.168.40.152:9092 (id: 2147483646 rack: null)
+14 May 2020 11:46:43,149 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions []
+14 May 2020 11:46:43,149 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 11:46:46,953 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 19
+14 May 2020 11:46:46,956 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [PROXY-EVENT-LOG-0]
+14 May 2020 11:46:46,957 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic PROXY-EVENT-LOG - partition 0 assigned.
+14 May 2020 11:51:42,863 INFO [kafka-producer-network-thread | producer-1] (org.apache.kafka.clients.Metadata.update:285) - Cluster ID: gsZT8xNUSViLGeSpG1neHg
+14 May 2020 12:02:55,605 INFO [lifecycleSupervisor-1-0] (org.apache.flume.node.PollingPropertiesFileConfigurationProvider.start:62) - Configuration provider starting
+14 May 2020 12:02:55,618 INFO [conf-file-poller-0] (org.apache.flume.node.PollingPropertiesFileConfigurationProvider$FileWatcherRunnable.run:138) - Reloading configuration file:/home/apache-flume-1.9.0-bin/conf/proxy/proxy_k2f.properties
+14 May 2020 12:02:55,622 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 12:02:55,623 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 12:02:55,624 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 12:02:55,624 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 12:02:55,624 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 12:02:55,624 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 12:02:55,624 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 12:02:55,624 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 12:02:55,625 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 12:02:55,625 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 12:02:55,625 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1117) - Added sinks: k1 Agent: proxy
+14 May 2020 12:02:55,625 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 12:02:55,626 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 12:02:55,626 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 12:02:55,627 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 12:02:55,627 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 12:02:55,628 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 12:02:55,628 WARN [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateConfigFilterSet:623) - Agent configuration for 'proxy' has no configfilters.
+14 May 2020 12:02:55,638 WARN [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateSources:702) - Agent configuration for 'proxy' has no sources.
+14 May 2020 12:02:55,641 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration.validateConfiguration:163) - Post-validation flume configuration contains configuration for agents: [proxy]
+14 May 2020 12:02:55,641 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:151) - Creating channels
+14 May 2020 12:02:55,647 INFO [conf-file-poller-0] (org.apache.flume.channel.DefaultChannelFactory.create:42) - Creating instance of channel c1 type org.apache.flume.channel.kafka.KafkaChannel
+14 May 2020 12:02:55,659 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:205) - Created channel c1
+14 May 2020 12:02:55,660 INFO [conf-file-poller-0] (org.apache.flume.sink.DefaultSinkFactory.create:42) - Creating instance of sink: k1, type: file_roll
+14 May 2020 12:02:55,666 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.getConfiguration:120) - Channel c1 connected to [k1]
+14 May 2020 12:02:55,672 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:162) - Starting new configuration:{ sourceRunners:{} sinkRunners:{k1=SinkRunner: { policy:org.apache.flume.sink.DefaultSinkProcessor@aba23f8 counterGroup:{ name:null counters:{} } }} channels:{c1=org.apache.flume.channel.kafka.KafkaChannel{name: c1}} }
+14 May 2020 12:02:55,673 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:169) - Starting Channel c1
+14 May 2020 12:02:55,674 INFO [lifecycleSupervisor-1-0] (org.apache.flume.channel.kafka.KafkaChannel.start:135) - Starting Kafka Channel: c1
+14 May 2020 12:02:55,713 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.config.AbstractConfig.logAll:279) - ProducerConfig values:
+ acks = all
+ batch.size = 16384
+ bootstrap.servers = [192.168.40.152:9092]
+ buffer.memory = 33554432
+ client.id =
+ compression.type = none
+ connections.max.idle.ms = 540000
+ enable.idempotence = false
+ interceptor.classes = []
+ key.serializer = class org.apache.kafka.common.serialization.StringSerializer
+ linger.ms = 0
+ max.block.ms = 60000
+ max.in.flight.requests.per.connection = 5
+ max.request.size = 1048576
+ metadata.max.age.ms = 300000
+ metric.reporters = []
+ metrics.num.samples = 2
+ metrics.recording.level = INFO
+ metrics.sample.window.ms = 30000
+ partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
+ receive.buffer.bytes = 32768
+ reconnect.backoff.max.ms = 1000
+ reconnect.backoff.ms = 50
+ request.timeout.ms = 30000
+ retries = 0
+ retry.backoff.ms = 100
+ sasl.client.callback.handler.class = null
+ sasl.jaas.config = null
+ sasl.kerberos.kinit.cmd = /usr/bin/kinit
+ sasl.kerberos.min.time.before.relogin = 60000
+ sasl.kerberos.service.name = null
+ sasl.kerberos.ticket.renew.jitter = 0.05
+ sasl.kerberos.ticket.renew.window.factor = 0.8
+ sasl.login.callback.handler.class = null
+ sasl.login.class = null
+ sasl.login.refresh.buffer.seconds = 300
+ sasl.login.refresh.min.period.seconds = 60
+ sasl.login.refresh.window.factor = 0.8
+ sasl.login.refresh.window.jitter = 0.05
+ sasl.mechanism = GSSAPI
+ security.protocol = PLAINTEXT
+ send.buffer.bytes = 131072
+ ssl.cipher.suites = null
+ ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
+ ssl.endpoint.identification.algorithm = https
+ ssl.key.password = null
+ ssl.keymanager.algorithm = SunX509
+ ssl.keystore.location = null
+ ssl.keystore.password = null
+ ssl.keystore.type = JKS
+ ssl.protocol = TLS
+ ssl.provider = null
+ ssl.secure.random.implementation = null
+ ssl.trustmanager.algorithm = PKIX
+ ssl.truststore.location = null
+ ssl.truststore.password = null
+ ssl.truststore.type = JKS
+ transaction.timeout.ms = 60000
+ transactional.id = null
+ value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
+
+14 May 2020 12:02:55,821 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:109) - Kafka version : 2.0.1
+14 May 2020 12:02:55,821 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:110) - Kafka commitId : fa14705e51bd2ce5
+14 May 2020 12:02:55,824 INFO [lifecycleSupervisor-1-0] (org.apache.flume.channel.kafka.KafkaChannel.start:143) - Topic = PROXY-EVENT-LOG
+14 May 2020 12:02:55,825 INFO [lifecycleSupervisor-1-0] (org.apache.flume.instrumentation.MonitoredCounterGroup.register:119) - Monitored counter group for type: CHANNEL, name: c1: Successfully registered new MBean.
+14 May 2020 12:02:55,826 INFO [lifecycleSupervisor-1-0] (org.apache.flume.instrumentation.MonitoredCounterGroup.start:95) - Component type: CHANNEL, name: c1 started
+14 May 2020 12:02:55,826 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:196) - Starting Sink k1
+14 May 2020 12:02:55,827 INFO [lifecycleSupervisor-1-1] (org.apache.flume.sink.RollingFileSink.start:111) - Starting org.apache.flume.sink.RollingFileSink{name:k1, channel:c1}...
+14 May 2020 12:02:55,829 INFO [lifecycleSupervisor-1-1] (org.apache.flume.instrumentation.MonitoredCounterGroup.register:119) - Monitored counter group for type: SINK, name: k1: Successfully registered new MBean.
+14 May 2020 12:02:55,829 INFO [lifecycleSupervisor-1-1] (org.apache.flume.instrumentation.MonitoredCounterGroup.start:95) - Component type: SINK, name: k1 started
+14 May 2020 12:02:55,831 INFO [lifecycleSupervisor-1-1] (org.apache.flume.sink.RollingFileSink.start:143) - RollingFileSink k1 started.
+14 May 2020 12:02:55,858 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.config.AbstractConfig.logAll:279) - ConsumerConfig values:
+ auto.commit.interval.ms = 5000
+ auto.offset.reset = latest
+ bootstrap.servers = [192.168.40.152:9092]
+ check.crcs = true
+ client.id =
+ connections.max.idle.ms = 540000
+ default.api.timeout.ms = 60000
+ enable.auto.commit = false
+ exclude.internal.topics = true
+ fetch.max.bytes = 52428800
+ fetch.max.wait.ms = 1000
+ fetch.min.bytes = 1048576
+ group.id = lxk_0514
+ heartbeat.interval.ms = 3000
+ interceptor.classes = []
+ internal.leave.group.on.close = true
+ isolation.level = read_uncommitted
+ key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
+ max.partition.fetch.bytes = 1048576
+ max.poll.interval.ms = 300000
+ max.poll.records = 500
+ metadata.max.age.ms = 300000
+ metric.reporters = []
+ metrics.num.samples = 2
+ metrics.recording.level = INFO
+ metrics.sample.window.ms = 30000
+ partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
+ receive.buffer.bytes = 65536
+ reconnect.backoff.max.ms = 1000
+ reconnect.backoff.ms = 50
+ request.timeout.ms = 30000
+ retry.backoff.ms = 100
+ sasl.client.callback.handler.class = null
+ sasl.jaas.config = null
+ sasl.kerberos.kinit.cmd = /usr/bin/kinit
+ sasl.kerberos.min.time.before.relogin = 60000
+ sasl.kerberos.service.name = null
+ sasl.kerberos.ticket.renew.jitter = 0.05
+ sasl.kerberos.ticket.renew.window.factor = 0.8
+ sasl.login.callback.handler.class = null
+ sasl.login.class = null
+ sasl.login.refresh.buffer.seconds = 300
+ sasl.login.refresh.min.period.seconds = 60
+ sasl.login.refresh.window.factor = 0.8
+ sasl.login.refresh.window.jitter = 0.05
+ sasl.mechanism = GSSAPI
+ security.protocol = PLAINTEXT
+ send.buffer.bytes = 131072
+ session.timeout.ms = 10000
+ ssl.cipher.suites = null
+ ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
+ ssl.endpoint.identification.algorithm = https
+ ssl.key.password = null
+ ssl.keymanager.algorithm = SunX509
+ ssl.keystore.location = null
+ ssl.keystore.password = null
+ ssl.keystore.type = JKS
+ ssl.protocol = TLS
+ ssl.provider = null
+ ssl.secure.random.implementation = null
+ ssl.trustmanager.algorithm = PKIX
+ ssl.truststore.location = null
+ ssl.truststore.password = null
+ ssl.truststore.type = JKS
+ value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
+
+14 May 2020 12:02:55,909 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:109) - Kafka version : 2.0.1
+14 May 2020 12:02:55,909 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:110) - Kafka commitId : fa14705e51bd2ce5
+14 May 2020 12:02:55,912 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.KafkaChannel.createConsumerAndRecords:304) - Created new consumer to connect to Kafka
+14 May 2020 12:02:56,156 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.Metadata.update:285) - Cluster ID: gsZT8xNUSViLGeSpG1neHg
+14 May 2020 12:02:56,244 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$FindCoordinatorResponseHandler.onSuccess:677) - [Consumer clientId=consumer-1, groupId=lxk_0514] Discovered group coordinator 192.168.40.152:9092 (id: 2147483646 rack: null)
+14 May 2020 12:02:56,248 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions []
+14 May 2020 12:02:56,249 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 12:02:59,926 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 21
+14 May 2020 12:02:59,928 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [PROXY-EVENT-LOG-0]
+14 May 2020 12:02:59,928 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic PROXY-EVENT-LOG - partition 0 assigned.
diff --git a/external-flume/conf/proxy/pro/2f2k.properties b/external-flume/conf_external/proxy/pro/2f2k.properties
index 5e5c573..5e5c573 100644
--- a/external-flume/conf/proxy/pro/2f2k.properties
+++ b/external-flume/conf_external/proxy/pro/2f2k.properties
diff --git a/external-flume/conf/proxy/pro/4f2k.properties b/external-flume/conf_external/proxy/pro/4f2k.properties
index bf0408b..bf0408b 100644
--- a/external-flume/conf/proxy/pro/4f2k.properties
+++ b/external-flume/conf_external/proxy/pro/4f2k.properties
diff --git a/external-flume/conf/proxy/pro/f2k.properties b/external-flume/conf_external/proxy/pro/f2k.properties
index 7047e42..7047e42 100644
--- a/external-flume/conf/proxy/pro/f2k.properties
+++ b/external-flume/conf_external/proxy/pro/f2k.properties
diff --git a/external-flume/conf/proxy/pro/f2k1.properties b/external-flume/conf_external/proxy/pro/f2k1.properties
index 2dadced..2dadced 100644
--- a/external-flume/conf/proxy/pro/f2k1.properties
+++ b/external-flume/conf_external/proxy/pro/f2k1.properties
diff --git a/external-flume/conf/proxy/pro/f2k2.properties b/external-flume/conf_external/proxy/pro/f2k2.properties
index e17c3e2..e17c3e2 100644
--- a/external-flume/conf/proxy/pro/f2k2.properties
+++ b/external-flume/conf_external/proxy/pro/f2k2.properties
diff --git a/external-flume/conf/proxy/pro/f2k3.properties b/external-flume/conf_external/proxy/pro/f2k3.properties
index 2dadced..2dadced 100644
--- a/external-flume/conf/proxy/pro/f2k3.properties
+++ b/external-flume/conf_external/proxy/pro/f2k3.properties
diff --git a/external-flume/conf/proxy/pro/file.properties b/external-flume/conf_external/proxy/pro/file.properties
index f6813e7..f6813e7 100644
--- a/external-flume/conf/proxy/pro/file.properties
+++ b/external-flume/conf_external/proxy/pro/file.properties
diff --git a/external-flume/conf/proxy/pro/file1.properties b/external-flume/conf_external/proxy/pro/file1.properties
index 17dac7f..17dac7f 100644
--- a/external-flume/conf/proxy/pro/file1.properties
+++ b/external-flume/conf_external/proxy/pro/file1.properties
diff --git a/external-flume/conf/proxy/pro/file2.properties b/external-flume/conf_external/proxy/pro/file2.properties
index f6813e7..f6813e7 100644
--- a/external-flume/conf/proxy/pro/file2.properties
+++ b/external-flume/conf_external/proxy/pro/file2.properties
diff --git a/external-flume/conf/proxy/pro/k2f.properties b/external-flume/conf_external/proxy/pro/k2f.properties
index 35d7d68..35d7d68 100644
--- a/external-flume/conf/proxy/pro/k2f.properties
+++ b/external-flume/conf_external/proxy/pro/k2f.properties
diff --git a/external-flume/conf/proxy/pro/kafka.properties b/external-flume/conf_external/proxy/pro/kafka.properties
index da560b2..da560b2 100644
--- a/external-flume/conf/proxy/pro/kafka.properties
+++ b/external-flume/conf_external/proxy/pro/kafka.properties
diff --git a/external-flume/conf/proxy/pro/kafka1.properties b/external-flume/conf_external/proxy/pro/kafka1.properties
index ed30b16..ed30b16 100644
--- a/external-flume/conf/proxy/pro/kafka1.properties
+++ b/external-flume/conf_external/proxy/pro/kafka1.properties
diff --git a/external-flume/conf/proxy/pro/kafka2.properties b/external-flume/conf_external/proxy/pro/kafka2.properties
index 6040a05..6040a05 100644
--- a/external-flume/conf/proxy/pro/kafka2.properties
+++ b/external-flume/conf_external/proxy/pro/kafka2.properties
diff --git a/external-flume/conf/proxy/pro/log4j.properties b/external-flume/conf_external/proxy/pro/log4j.properties
index c948bff..c948bff 100644
--- a/external-flume/conf/proxy/pro/log4j.properties
+++ b/external-flume/conf_external/proxy/pro/log4j.properties
diff --git a/external-flume/conf/proxy/pro/tail.properties b/external-flume/conf_external/proxy/pro/tail.properties
index c677318..c677318 100644
--- a/external-flume/conf/proxy/pro/tail.properties
+++ b/external-flume/conf_external/proxy/pro/tail.properties
diff --git a/external-flume/conf/proxy/proxy_k2f.properties b/external-flume/conf_external/proxy/proxy_k2f.properties
index 155b433..155b433 100644
--- a/external-flume/conf/proxy/proxy_k2f.properties
+++ b/external-flume/conf_external/proxy/proxy_k2f.properties
diff --git a/external-flume/conf_external/proxy/restart_log/restart_proxy_k2f.log b/external-flume/conf_external/proxy/restart_log/restart_proxy_k2f.log
new file mode 100644
index 0000000..a6998d2
--- /dev/null
+++ b/external-flume/conf_external/proxy/restart_log/restart_proxy_k2f.log
@@ -0,0 +1,5 @@
+##############################################################
+20200514112452 ---> the proxy_k2f_APP restart ---> 1
+##############################################################
+20200514114642 ---> the proxy_k2f_APP restart ---> 1
+20200514120255 ---> the proxy_k2f_APP restart ---> 1
diff --git a/external-flume/conf/radius/start_flume.sh b/external-flume/conf_external/proxy/start_flume.sh
index 01dde8b..01dde8b 100755
--- a/external-flume/conf/radius/start_flume.sh
+++ b/external-flume/conf_external/proxy/start_flume.sh
diff --git a/external-flume/conf/proxy/stop_flume.sh b/external-flume/conf_external/proxy/stop_flume.sh
index e4417aa..e4417aa 100755
--- a/external-flume/conf/proxy/stop_flume.sh
+++ b/external-flume/conf_external/proxy/stop_flume.sh
diff --git a/external-flume/conf/security/clean_start_flume.sh b/external-flume/conf_external/security/clean_start_flume.sh
index ef811a5..ef811a5 100755
--- a/external-flume/conf/security/clean_start_flume.sh
+++ b/external-flume/conf_external/security/clean_start_flume.sh
diff --git a/external-flume/conf/radius/count_flume.sh b/external-flume/conf_external/security/count_flume.sh
index cc8452b..cc8452b 100755
--- a/external-flume/conf/radius/count_flume.sh
+++ b/external-flume/conf_external/security/count_flume.sh
diff --git a/external-flume/conf/security/dae_security.sh b/external-flume/conf_external/security/dae_security.sh
index cf0e1f0..cf0e1f0 100755
--- a/external-flume/conf/security/dae_security.sh
+++ b/external-flume/conf_external/security/dae_security.sh
diff --git a/external-flume/conf/security/flume-env.sh b/external-flume/conf_external/security/flume-env.sh
index 63979aa..63979aa 100755
--- a/external-flume/conf/security/flume-env.sh
+++ b/external-flume/conf_external/security/flume-env.sh
diff --git a/external-flume/conf/radius/log4j.properties b/external-flume/conf_external/security/log4j.properties
index c948bff..c948bff 100644
--- a/external-flume/conf/radius/log4j.properties
+++ b/external-flume/conf_external/security/log4j.properties
diff --git a/external-flume/conf_external/security/logs/flume.log b/external-flume/conf_external/security/logs/flume.log
new file mode 100644
index 0000000..22cde2c
--- /dev/null
+++ b/external-flume/conf_external/security/logs/flume.log
@@ -0,0 +1,189 @@
+14 May 2020 11:46:42,542 INFO [lifecycleSupervisor-1-0] (org.apache.flume.node.PollingPropertiesFileConfigurationProvider.start:62) - Configuration provider starting
+14 May 2020 11:46:42,551 INFO [conf-file-poller-0] (org.apache.flume.node.PollingPropertiesFileConfigurationProvider$FileWatcherRunnable.run:138) - Reloading configuration file:/home/apache-flume-1.9.0-bin/conf/security/security_k2f.properties
+14 May 2020 11:46:42,556 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,557 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1117) - Added sinks: k1 Agent: security
+14 May 2020 11:46:42,557 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,557 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,557 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,558 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,558 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,558 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,558 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,558 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,558 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,559 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,559 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,559 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,560 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,561 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,561 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,561 WARN [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateConfigFilterSet:623) - Agent configuration for 'security' has no configfilters.
+14 May 2020 11:46:42,571 WARN [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateSources:702) - Agent configuration for 'security' has no sources.
+14 May 2020 11:46:42,573 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration.validateConfiguration:163) - Post-validation flume configuration contains configuration for agents: [security]
+14 May 2020 11:46:42,574 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:151) - Creating channels
+14 May 2020 11:46:42,579 INFO [conf-file-poller-0] (org.apache.flume.channel.DefaultChannelFactory.create:42) - Creating instance of channel c1 type org.apache.flume.channel.kafka.KafkaChannel
+14 May 2020 11:46:42,591 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:205) - Created channel c1
+14 May 2020 11:46:42,592 INFO [conf-file-poller-0] (org.apache.flume.sink.DefaultSinkFactory.create:42) - Creating instance of sink: k1, type: file_roll
+14 May 2020 11:46:42,598 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.getConfiguration:120) - Channel c1 connected to [k1]
+14 May 2020 11:46:42,604 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:162) - Starting new configuration:{ sourceRunners:{} sinkRunners:{k1=SinkRunner: { policy:org.apache.flume.sink.DefaultSinkProcessor@7e8b5303 counterGroup:{ name:null counters:{} } }} channels:{c1=org.apache.flume.channel.kafka.KafkaChannel{name: c1}} }
+14 May 2020 11:46:42,605 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:169) - Starting Channel c1
+14 May 2020 11:46:42,605 INFO [lifecycleSupervisor-1-0] (org.apache.flume.channel.kafka.KafkaChannel.start:135) - Starting Kafka Channel: c1
+14 May 2020 11:46:42,636 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.config.AbstractConfig.logAll:279) - ProducerConfig values:
+ acks = all
+ batch.size = 16384
+ bootstrap.servers = [192.168.40.152:9092]
+ buffer.memory = 33554432
+ client.id =
+ compression.type = none
+ connections.max.idle.ms = 540000
+ enable.idempotence = false
+ interceptor.classes = []
+ key.serializer = class org.apache.kafka.common.serialization.StringSerializer
+ linger.ms = 0
+ max.block.ms = 60000
+ max.in.flight.requests.per.connection = 5
+ max.request.size = 1048576
+ metadata.max.age.ms = 300000
+ metric.reporters = []
+ metrics.num.samples = 2
+ metrics.recording.level = INFO
+ metrics.sample.window.ms = 30000
+ partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
+ receive.buffer.bytes = 32768
+ reconnect.backoff.max.ms = 1000
+ reconnect.backoff.ms = 50
+ request.timeout.ms = 30000
+ retries = 0
+ retry.backoff.ms = 100
+ sasl.client.callback.handler.class = null
+ sasl.jaas.config = null
+ sasl.kerberos.kinit.cmd = /usr/bin/kinit
+ sasl.kerberos.min.time.before.relogin = 60000
+ sasl.kerberos.service.name = null
+ sasl.kerberos.ticket.renew.jitter = 0.05
+ sasl.kerberos.ticket.renew.window.factor = 0.8
+ sasl.login.callback.handler.class = null
+ sasl.login.class = null
+ sasl.login.refresh.buffer.seconds = 300
+ sasl.login.refresh.min.period.seconds = 60
+ sasl.login.refresh.window.factor = 0.8
+ sasl.login.refresh.window.jitter = 0.05
+ sasl.mechanism = GSSAPI
+ security.protocol = PLAINTEXT
+ send.buffer.bytes = 131072
+ ssl.cipher.suites = null
+ ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
+ ssl.endpoint.identification.algorithm = https
+ ssl.key.password = null
+ ssl.keymanager.algorithm = SunX509
+ ssl.keystore.location = null
+ ssl.keystore.password = null
+ ssl.keystore.type = JKS
+ ssl.protocol = TLS
+ ssl.provider = null
+ ssl.secure.random.implementation = null
+ ssl.trustmanager.algorithm = PKIX
+ ssl.truststore.location = null
+ ssl.truststore.password = null
+ ssl.truststore.type = JKS
+ transaction.timeout.ms = 60000
+ transactional.id = null
+ value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
+
+14 May 2020 11:46:42,692 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:109) - Kafka version : 2.0.1
+14 May 2020 11:46:42,692 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:110) - Kafka commitId : fa14705e51bd2ce5
+14 May 2020 11:46:42,694 INFO [lifecycleSupervisor-1-0] (org.apache.flume.channel.kafka.KafkaChannel.start:143) - Topic = SECURITY-EVENT-LOG
+14 May 2020 11:46:42,695 INFO [lifecycleSupervisor-1-0] (org.apache.flume.instrumentation.MonitoredCounterGroup.register:119) - Monitored counter group for type: CHANNEL, name: c1: Successfully registered new MBean.
+14 May 2020 11:46:42,695 INFO [lifecycleSupervisor-1-0] (org.apache.flume.instrumentation.MonitoredCounterGroup.start:95) - Component type: CHANNEL, name: c1 started
+14 May 2020 11:46:42,695 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:196) - Starting Sink k1
+14 May 2020 11:46:42,697 INFO [lifecycleSupervisor-1-1] (org.apache.flume.sink.RollingFileSink.start:111) - Starting org.apache.flume.sink.RollingFileSink{name:k1, channel:c1}...
+14 May 2020 11:46:42,698 INFO [lifecycleSupervisor-1-1] (org.apache.flume.instrumentation.MonitoredCounterGroup.register:119) - Monitored counter group for type: SINK, name: k1: Successfully registered new MBean.
+14 May 2020 11:46:42,698 INFO [lifecycleSupervisor-1-1] (org.apache.flume.instrumentation.MonitoredCounterGroup.start:95) - Component type: SINK, name: k1 started
+14 May 2020 11:46:42,700 INFO [lifecycleSupervisor-1-1] (org.apache.flume.sink.RollingFileSink.start:143) - RollingFileSink k1 started.
+14 May 2020 11:46:42,734 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.config.AbstractConfig.logAll:279) - ConsumerConfig values:
+ auto.commit.interval.ms = 5000
+ auto.offset.reset = latest
+ bootstrap.servers = [192.168.40.152:9092]
+ check.crcs = true
+ client.id =
+ connections.max.idle.ms = 540000
+ default.api.timeout.ms = 60000
+ enable.auto.commit = false
+ exclude.internal.topics = true
+ fetch.max.bytes = 52428800
+ fetch.max.wait.ms = 1000
+ fetch.min.bytes = 1048576
+ group.id = lxk_0514
+ heartbeat.interval.ms = 3000
+ interceptor.classes = []
+ internal.leave.group.on.close = true
+ isolation.level = read_uncommitted
+ key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
+ max.partition.fetch.bytes = 1048576
+ max.poll.interval.ms = 300000
+ max.poll.records = 500
+ metadata.max.age.ms = 300000
+ metric.reporters = []
+ metrics.num.samples = 2
+ metrics.recording.level = INFO
+ metrics.sample.window.ms = 30000
+ partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
+ receive.buffer.bytes = 65536
+ reconnect.backoff.max.ms = 1000
+ reconnect.backoff.ms = 50
+ request.timeout.ms = 30000
+ retry.backoff.ms = 100
+ sasl.client.callback.handler.class = null
+ sasl.jaas.config = null
+ sasl.kerberos.kinit.cmd = /usr/bin/kinit
+ sasl.kerberos.min.time.before.relogin = 60000
+ sasl.kerberos.service.name = null
+ sasl.kerberos.ticket.renew.jitter = 0.05
+ sasl.kerberos.ticket.renew.window.factor = 0.8
+ sasl.login.callback.handler.class = null
+ sasl.login.class = null
+ sasl.login.refresh.buffer.seconds = 300
+ sasl.login.refresh.min.period.seconds = 60
+ sasl.login.refresh.window.factor = 0.8
+ sasl.login.refresh.window.jitter = 0.05
+ sasl.mechanism = GSSAPI
+ security.protocol = PLAINTEXT
+ send.buffer.bytes = 131072
+ session.timeout.ms = 10000
+ ssl.cipher.suites = null
+ ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
+ ssl.endpoint.identification.algorithm = https
+ ssl.key.password = null
+ ssl.keymanager.algorithm = SunX509
+ ssl.keystore.location = null
+ ssl.keystore.password = null
+ ssl.keystore.type = JKS
+ ssl.protocol = TLS
+ ssl.provider = null
+ ssl.secure.random.implementation = null
+ ssl.trustmanager.algorithm = PKIX
+ ssl.truststore.location = null
+ ssl.truststore.password = null
+ ssl.truststore.type = JKS
+ value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
+
+14 May 2020 11:46:42,796 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:109) - Kafka version : 2.0.1
+14 May 2020 11:46:42,796 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:110) - Kafka commitId : fa14705e51bd2ce5
+14 May 2020 11:46:42,800 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.KafkaChannel.createConsumerAndRecords:304) - Created new consumer to connect to Kafka
+14 May 2020 11:46:43,000 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.Metadata.update:285) - Cluster ID: gsZT8xNUSViLGeSpG1neHg
+14 May 2020 11:46:43,001 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$FindCoordinatorResponseHandler.onSuccess:677) - [Consumer clientId=consumer-1, groupId=lxk_0514] Discovered group coordinator 192.168.40.152:9092 (id: 2147483646 rack: null)
+14 May 2020 11:46:43,003 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions []
+14 May 2020 11:46:43,003 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 11:46:46,812 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 19
+14 May 2020 11:46:46,814 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [SECURITY-EVENT-LOG-0]
+14 May 2020 11:46:46,815 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic SECURITY-EVENT-LOG - partition 0 assigned.
+14 May 2020 11:51:42,694 INFO [kafka-producer-network-thread | producer-1] (org.apache.kafka.clients.Metadata.update:285) - Cluster ID: gsZT8xNUSViLGeSpG1neHg
+14 May 2020 12:02:49,163 INFO [kafka-coordinator-heartbeat-thread | lxk_0514] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$HeartbeatResponseHandler.handle:855) - [Consumer clientId=consumer-1, groupId=lxk_0514] Attempt to heartbeat failed since group is rebalancing
+14 May 2020 12:02:52,169 INFO [kafka-coordinator-heartbeat-thread | lxk_0514] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$HeartbeatResponseHandler.handle:855) - [Consumer clientId=consumer-1, groupId=lxk_0514] Attempt to heartbeat failed since group is rebalancing
+14 May 2020 12:02:52,762 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions [SECURITY-EVENT-LOG-0]
+14 May 2020 12:02:52,763 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsRevoked:785) - topic SECURITY-EVENT-LOG - partition 0 revoked.
+14 May 2020 12:02:52,763 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 12:02:58,267 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 12:03:03,769 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 21
+14 May 2020 12:03:03,770 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [SECURITY-EVENT-LOG-0]
+14 May 2020 12:03:03,771 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic SECURITY-EVENT-LOG - partition 0 assigned.
diff --git a/external-flume/conf/radius/pro/2f2k.properties b/external-flume/conf_external/security/pro/2f2k.properties
index 5e5c573..5e5c573 100644
--- a/external-flume/conf/radius/pro/2f2k.properties
+++ b/external-flume/conf_external/security/pro/2f2k.properties
diff --git a/external-flume/conf/radius/pro/4f2k.properties b/external-flume/conf_external/security/pro/4f2k.properties
index bf0408b..bf0408b 100644
--- a/external-flume/conf/radius/pro/4f2k.properties
+++ b/external-flume/conf_external/security/pro/4f2k.properties
diff --git a/external-flume/conf/radius/pro/f2k.properties b/external-flume/conf_external/security/pro/f2k.properties
index 7047e42..7047e42 100644
--- a/external-flume/conf/radius/pro/f2k.properties
+++ b/external-flume/conf_external/security/pro/f2k.properties
diff --git a/external-flume/conf/radius/pro/f2k1.properties b/external-flume/conf_external/security/pro/f2k1.properties
index 2dadced..2dadced 100644
--- a/external-flume/conf/radius/pro/f2k1.properties
+++ b/external-flume/conf_external/security/pro/f2k1.properties
diff --git a/external-flume/conf/radius/pro/f2k2.properties b/external-flume/conf_external/security/pro/f2k2.properties
index e17c3e2..e17c3e2 100644
--- a/external-flume/conf/radius/pro/f2k2.properties
+++ b/external-flume/conf_external/security/pro/f2k2.properties
diff --git a/external-flume/conf/radius/pro/f2k3.properties b/external-flume/conf_external/security/pro/f2k3.properties
index 2dadced..2dadced 100644
--- a/external-flume/conf/radius/pro/f2k3.properties
+++ b/external-flume/conf_external/security/pro/f2k3.properties
diff --git a/external-flume/conf/radius/pro/file.properties b/external-flume/conf_external/security/pro/file.properties
index f6813e7..f6813e7 100644
--- a/external-flume/conf/radius/pro/file.properties
+++ b/external-flume/conf_external/security/pro/file.properties
diff --git a/external-flume/conf/radius/pro/file1.properties b/external-flume/conf_external/security/pro/file1.properties
index 17dac7f..17dac7f 100644
--- a/external-flume/conf/radius/pro/file1.properties
+++ b/external-flume/conf_external/security/pro/file1.properties
diff --git a/external-flume/conf/radius/pro/file2.properties b/external-flume/conf_external/security/pro/file2.properties
index f6813e7..f6813e7 100644
--- a/external-flume/conf/radius/pro/file2.properties
+++ b/external-flume/conf_external/security/pro/file2.properties
diff --git a/external-flume/conf/radius/pro/k2f.properties b/external-flume/conf_external/security/pro/k2f.properties
index 35d7d68..35d7d68 100644
--- a/external-flume/conf/radius/pro/k2f.properties
+++ b/external-flume/conf_external/security/pro/k2f.properties
diff --git a/external-flume/conf/radius/pro/kafka.properties b/external-flume/conf_external/security/pro/kafka.properties
index da560b2..da560b2 100644
--- a/external-flume/conf/radius/pro/kafka.properties
+++ b/external-flume/conf_external/security/pro/kafka.properties
diff --git a/external-flume/conf/radius/pro/kafka1.properties b/external-flume/conf_external/security/pro/kafka1.properties
index ed30b16..ed30b16 100644
--- a/external-flume/conf/radius/pro/kafka1.properties
+++ b/external-flume/conf_external/security/pro/kafka1.properties
diff --git a/external-flume/conf/radius/pro/kafka2.properties b/external-flume/conf_external/security/pro/kafka2.properties
index 6040a05..6040a05 100644
--- a/external-flume/conf/radius/pro/kafka2.properties
+++ b/external-flume/conf_external/security/pro/kafka2.properties
diff --git a/external-flume/conf/radius/pro/log4j.properties b/external-flume/conf_external/security/pro/log4j.properties
index c948bff..c948bff 100644
--- a/external-flume/conf/radius/pro/log4j.properties
+++ b/external-flume/conf_external/security/pro/log4j.properties
diff --git a/external-flume/conf/radius/pro/tail.properties b/external-flume/conf_external/security/pro/tail.properties
index c677318..c677318 100644
--- a/external-flume/conf/radius/pro/tail.properties
+++ b/external-flume/conf_external/security/pro/tail.properties
diff --git a/external-flume/conf/connection/restart_log/restart_connection_k2f.log b/external-flume/conf_external/security/restart_log/restart_connection_k2f.log
index 281d596..281d596 100644
--- a/external-flume/conf/connection/restart_log/restart_connection_k2f.log
+++ b/external-flume/conf_external/security/restart_log/restart_connection_k2f.log
diff --git a/external-flume/conf/radius/restart_log/restart_f2k.log b/external-flume/conf_external/security/restart_log/restart_f2k.log
index 9d5fe99..9d5fe99 100644
--- a/external-flume/conf/radius/restart_log/restart_f2k.log
+++ b/external-flume/conf_external/security/restart_log/restart_f2k.log
diff --git a/external-flume/conf/radius/restart_log/restart_security_f2k.log b/external-flume/conf_external/security/restart_log/restart_security_f2k.log
index 690385d..690385d 100644
--- a/external-flume/conf/radius/restart_log/restart_security_f2k.log
+++ b/external-flume/conf_external/security/restart_log/restart_security_f2k.log
diff --git a/external-flume/conf_external/security/restart_log/restart_security_k2f.log b/external-flume/conf_external/security/restart_log/restart_security_k2f.log
new file mode 100644
index 0000000..70f25c1
--- /dev/null
+++ b/external-flume/conf_external/security/restart_log/restart_security_k2f.log
@@ -0,0 +1,4 @@
+##############################################################
+20200514112452 ---> the security_k2f_APP restart ---> 1
+##############################################################
+20200514114642 ---> the security_k2f_APP restart ---> 1
diff --git a/external-flume/conf/security/security_k2f.properties b/external-flume/conf_external/security/security_k2f.properties
index b60c2e7..b60c2e7 100644
--- a/external-flume/conf/security/security_k2f.properties
+++ b/external-flume/conf_external/security/security_k2f.properties
diff --git a/external-flume/conf/security/start_flume.sh b/external-flume/conf_external/security/start_flume.sh
index 01dde8b..01dde8b 100755
--- a/external-flume/conf/security/start_flume.sh
+++ b/external-flume/conf_external/security/start_flume.sh
diff --git a/external-flume/conf/radius/stop_flume.sh b/external-flume/conf_external/security/stop_flume.sh
index e4417aa..e4417aa 100755
--- a/external-flume/conf/radius/stop_flume.sh
+++ b/external-flume/conf_external/security/stop_flume.sh
diff --git a/internal-flume/conf/connection/clean_start_flume.sh b/external-flume/conf_external/traffic/clean_start_flume.sh
index ef811a5..ef811a5 100755
--- a/internal-flume/conf/connection/clean_start_flume.sh
+++ b/external-flume/conf_external/traffic/clean_start_flume.sh
diff --git a/external-flume/conf/security/count_flume.sh b/external-flume/conf_external/traffic/count_flume.sh
index cc8452b..cc8452b 100755
--- a/external-flume/conf/security/count_flume.sh
+++ b/external-flume/conf_external/traffic/count_flume.sh
diff --git a/external-flume/conf/radius/dae_radius.sh b/external-flume/conf_external/traffic/dae_connection.sh
index 04563ce..b7ba0d9 100755
--- a/external-flume/conf/radius/dae_radius.sh
+++ b/external-flume/conf_external/traffic/dae_connection.sh
@@ -1,9 +1,9 @@
#!/bin/sh
#JAR_NAME=ktk_ip_asn.properties
-PROPERTIES_NAME=radius_k2f
+PROPERTIES_NAME=connection_k2f
#flume进程名称
-FLUME_NAME=radius
+FLUME_NAME=connection
#flume根目录
BASE_DIR=$(cd $(dirname $0); cd ../../; pwd)
#当前路径
diff --git a/internal-flume/conf/connection/flume-env.sh b/external-flume/conf_external/traffic/flume-env.sh
index 63979aa..63979aa 100755
--- a/internal-flume/conf/connection/flume-env.sh
+++ b/external-flume/conf_external/traffic/flume-env.sh
diff --git a/external-flume/conf/security/log4j.properties b/external-flume/conf_external/traffic/log4j.properties
index c948bff..c948bff 100644
--- a/external-flume/conf/security/log4j.properties
+++ b/external-flume/conf_external/traffic/log4j.properties
diff --git a/external-flume/conf_external/traffic/logs/flume.log b/external-flume/conf_external/traffic/logs/flume.log
new file mode 100644
index 0000000..cb74ece
--- /dev/null
+++ b/external-flume/conf_external/traffic/logs/flume.log
@@ -0,0 +1,201 @@
+14 May 2020 11:46:42,409 INFO [lifecycleSupervisor-1-0] (org.apache.flume.node.PollingPropertiesFileConfigurationProvider.start:62) - Configuration provider starting
+14 May 2020 11:46:42,421 INFO [conf-file-poller-0] (org.apache.flume.node.PollingPropertiesFileConfigurationProvider$FileWatcherRunnable.run:138) - Reloading configuration file:/home/apache-flume-1.9.0-bin/conf/connection/connection_k2f.properties
+14 May 2020 11:46:42,431 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,433 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,433 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,434 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,434 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,435 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,435 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1117) - Added sinks: k1 Agent: connection
+14 May 2020 11:46:42,435 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,436 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,436 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,436 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,437 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,437 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:k1
+14 May 2020 11:46:42,438 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,440 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,440 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,441 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addComponentConfig:1203) - Processing:c1
+14 May 2020 11:46:42,441 WARN [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateConfigFilterSet:623) - Agent configuration for 'connection' has no configfilters.
+14 May 2020 11:46:42,464 WARN [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateSources:702) - Agent configuration for 'connection' has no sources.
+14 May 2020 11:46:42,469 INFO [conf-file-poller-0] (org.apache.flume.conf.FlumeConfiguration.validateConfiguration:163) - Post-validation flume configuration contains configuration for agents: [connection]
+14 May 2020 11:46:42,469 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:151) - Creating channels
+14 May 2020 11:46:42,481 INFO [conf-file-poller-0] (org.apache.flume.channel.DefaultChannelFactory.create:42) - Creating instance of channel c1 type org.apache.flume.channel.kafka.KafkaChannel
+14 May 2020 11:46:42,503 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:205) - Created channel c1
+14 May 2020 11:46:42,504 INFO [conf-file-poller-0] (org.apache.flume.sink.DefaultSinkFactory.create:42) - Creating instance of sink: k1, type: file_roll
+14 May 2020 11:46:42,514 INFO [conf-file-poller-0] (org.apache.flume.node.AbstractConfigurationProvider.getConfiguration:120) - Channel c1 connected to [k1]
+14 May 2020 11:46:42,524 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:162) - Starting new configuration:{ sourceRunners:{} sinkRunners:{k1=SinkRunner: { policy:org.apache.flume.sink.DefaultSinkProcessor@5cf2aeb7 counterGroup:{ name:null counters:{} } }} channels:{c1=org.apache.flume.channel.kafka.KafkaChannel{name: c1}} }
+14 May 2020 11:46:42,525 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:169) - Starting Channel c1
+14 May 2020 11:46:42,526 INFO [lifecycleSupervisor-1-0] (org.apache.flume.channel.kafka.KafkaChannel.start:135) - Starting Kafka Channel: c1
+14 May 2020 11:46:42,562 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.config.AbstractConfig.logAll:279) - ProducerConfig values:
+ acks = all
+ batch.size = 16384
+ bootstrap.servers = [192.168.40.152:9092]
+ buffer.memory = 33554432
+ client.id =
+ compression.type = none
+ connections.max.idle.ms = 540000
+ enable.idempotence = false
+ interceptor.classes = []
+ key.serializer = class org.apache.kafka.common.serialization.StringSerializer
+ linger.ms = 0
+ max.block.ms = 60000
+ max.in.flight.requests.per.connection = 5
+ max.request.size = 1048576
+ metadata.max.age.ms = 300000
+ metric.reporters = []
+ metrics.num.samples = 2
+ metrics.recording.level = INFO
+ metrics.sample.window.ms = 30000
+ partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
+ receive.buffer.bytes = 32768
+ reconnect.backoff.max.ms = 1000
+ reconnect.backoff.ms = 50
+ request.timeout.ms = 30000
+ retries = 0
+ retry.backoff.ms = 100
+ sasl.client.callback.handler.class = null
+ sasl.jaas.config = null
+ sasl.kerberos.kinit.cmd = /usr/bin/kinit
+ sasl.kerberos.min.time.before.relogin = 60000
+ sasl.kerberos.service.name = null
+ sasl.kerberos.ticket.renew.jitter = 0.05
+ sasl.kerberos.ticket.renew.window.factor = 0.8
+ sasl.login.callback.handler.class = null
+ sasl.login.class = null
+ sasl.login.refresh.buffer.seconds = 300
+ sasl.login.refresh.min.period.seconds = 60
+ sasl.login.refresh.window.factor = 0.8
+ sasl.login.refresh.window.jitter = 0.05
+ sasl.mechanism = GSSAPI
+ security.protocol = PLAINTEXT
+ send.buffer.bytes = 131072
+ ssl.cipher.suites = null
+ ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
+ ssl.endpoint.identification.algorithm = https
+ ssl.key.password = null
+ ssl.keymanager.algorithm = SunX509
+ ssl.keystore.location = null
+ ssl.keystore.password = null
+ ssl.keystore.type = JKS
+ ssl.protocol = TLS
+ ssl.provider = null
+ ssl.secure.random.implementation = null
+ ssl.trustmanager.algorithm = PKIX
+ ssl.truststore.location = null
+ ssl.truststore.password = null
+ ssl.truststore.type = JKS
+ transaction.timeout.ms = 60000
+ transactional.id = null
+ value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
+
+14 May 2020 11:46:42,638 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:109) - Kafka version : 2.0.1
+14 May 2020 11:46:42,639 INFO [lifecycleSupervisor-1-0] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:110) - Kafka commitId : fa14705e51bd2ce5
+14 May 2020 11:46:42,641 INFO [lifecycleSupervisor-1-0] (org.apache.flume.channel.kafka.KafkaChannel.start:143) - Topic = CONNECTION-RECORD-LOG
+14 May 2020 11:46:42,641 INFO [lifecycleSupervisor-1-0] (org.apache.flume.instrumentation.MonitoredCounterGroup.register:119) - Monitored counter group for type: CHANNEL, name: c1: Successfully registered new MBean.
+14 May 2020 11:46:42,642 INFO [lifecycleSupervisor-1-0] (org.apache.flume.instrumentation.MonitoredCounterGroup.start:95) - Component type: CHANNEL, name: c1 started
+14 May 2020 11:46:42,642 INFO [conf-file-poller-0] (org.apache.flume.node.Application.startAllComponents:196) - Starting Sink k1
+14 May 2020 11:46:42,643 INFO [lifecycleSupervisor-1-1] (org.apache.flume.sink.RollingFileSink.start:111) - Starting org.apache.flume.sink.RollingFileSink{name:k1, channel:c1}...
+14 May 2020 11:46:42,644 INFO [lifecycleSupervisor-1-1] (org.apache.flume.instrumentation.MonitoredCounterGroup.register:119) - Monitored counter group for type: SINK, name: k1: Successfully registered new MBean.
+14 May 2020 11:46:42,644 INFO [lifecycleSupervisor-1-1] (org.apache.flume.instrumentation.MonitoredCounterGroup.start:95) - Component type: SINK, name: k1 started
+14 May 2020 11:46:42,645 INFO [lifecycleSupervisor-1-1] (org.apache.flume.sink.RollingFileSink.start:143) - RollingFileSink k1 started.
+14 May 2020 11:46:42,661 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.config.AbstractConfig.logAll:279) - ConsumerConfig values:
+ auto.commit.interval.ms = 5000
+ auto.offset.reset = latest
+ bootstrap.servers = [192.168.40.152:9092]
+ check.crcs = true
+ client.id =
+ connections.max.idle.ms = 540000
+ default.api.timeout.ms = 60000
+ enable.auto.commit = false
+ exclude.internal.topics = true
+ fetch.max.bytes = 52428800
+ fetch.max.wait.ms = 1000
+ fetch.min.bytes = 1048576
+ group.id = lxk_0514
+ heartbeat.interval.ms = 3000
+ interceptor.classes = []
+ internal.leave.group.on.close = true
+ isolation.level = read_uncommitted
+ key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
+ max.partition.fetch.bytes = 1048576
+ max.poll.interval.ms = 300000
+ max.poll.records = 500
+ metadata.max.age.ms = 300000
+ metric.reporters = []
+ metrics.num.samples = 2
+ metrics.recording.level = INFO
+ metrics.sample.window.ms = 30000
+ partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
+ receive.buffer.bytes = 65536
+ reconnect.backoff.max.ms = 1000
+ reconnect.backoff.ms = 50
+ request.timeout.ms = 30000
+ retry.backoff.ms = 100
+ sasl.client.callback.handler.class = null
+ sasl.jaas.config = null
+ sasl.kerberos.kinit.cmd = /usr/bin/kinit
+ sasl.kerberos.min.time.before.relogin = 60000
+ sasl.kerberos.service.name = null
+ sasl.kerberos.ticket.renew.jitter = 0.05
+ sasl.kerberos.ticket.renew.window.factor = 0.8
+ sasl.login.callback.handler.class = null
+ sasl.login.class = null
+ sasl.login.refresh.buffer.seconds = 300
+ sasl.login.refresh.min.period.seconds = 60
+ sasl.login.refresh.window.factor = 0.8
+ sasl.login.refresh.window.jitter = 0.05
+ sasl.mechanism = GSSAPI
+ security.protocol = PLAINTEXT
+ send.buffer.bytes = 131072
+ session.timeout.ms = 10000
+ ssl.cipher.suites = null
+ ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
+ ssl.endpoint.identification.algorithm = https
+ ssl.key.password = null
+ ssl.keymanager.algorithm = SunX509
+ ssl.keystore.location = null
+ ssl.keystore.password = null
+ ssl.keystore.type = JKS
+ ssl.protocol = TLS
+ ssl.provider = null
+ ssl.secure.random.implementation = null
+ ssl.trustmanager.algorithm = PKIX
+ ssl.truststore.location = null
+ ssl.truststore.password = null
+ ssl.truststore.type = JKS
+ value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
+
+14 May 2020 11:46:42,696 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:109) - Kafka version : 2.0.1
+14 May 2020 11:46:42,697 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.common.utils.AppInfoParser$AppInfo.<init>:110) - Kafka commitId : fa14705e51bd2ce5
+14 May 2020 11:46:42,700 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.KafkaChannel.createConsumerAndRecords:304) - Created new consumer to connect to Kafka
+14 May 2020 11:46:42,870 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.Metadata.update:285) - Cluster ID: gsZT8xNUSViLGeSpG1neHg
+14 May 2020 11:46:42,871 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$FindCoordinatorResponseHandler.onSuccess:677) - [Consumer clientId=consumer-1, groupId=lxk_0514] Discovered group coordinator 192.168.40.152:9092 (id: 2147483646 rack: null)
+14 May 2020 11:46:42,873 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions []
+14 May 2020 11:46:42,873 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 11:46:42,887 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 18
+14 May 2020 11:46:42,888 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 11:46:42,889 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic CONNECTION-RECORD-LOG - partition 0 assigned.
+14 May 2020 11:46:45,901 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$HeartbeatResponseHandler.handle:855) - [Consumer clientId=consumer-1, groupId=lxk_0514] Attempt to heartbeat failed since group is rebalancing
+14 May 2020 11:46:45,910 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 11:46:45,910 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsRevoked:785) - topic CONNECTION-RECORD-LOG - partition 0 revoked.
+14 May 2020 11:46:45,911 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 11:46:45,921 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 19
+14 May 2020 11:46:45,922 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 11:46:45,923 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic CONNECTION-RECORD-LOG - partition 0 assigned.
+14 May 2020 11:51:42,641 INFO [kafka-producer-network-thread | producer-1] (org.apache.kafka.clients.Metadata.update:285) - Cluster ID: gsZT8xNUSViLGeSpG1neHg
+14 May 2020 12:02:48,119 INFO [kafka-coordinator-heartbeat-thread | lxk_0514] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$HeartbeatResponseHandler.handle:855) - [Consumer clientId=consumer-1, groupId=lxk_0514] Attempt to heartbeat failed since group is rebalancing
+14 May 2020 12:02:48,715 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 12:02:48,716 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsRevoked:785) - topic CONNECTION-RECORD-LOG - partition 0 revoked.
+14 May 2020 12:02:48,717 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 12:02:54,224 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 20
+14 May 2020 12:02:54,225 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 12:02:54,225 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic CONNECTION-RECORD-LOG - partition 0 assigned.
+14 May 2020 12:02:57,326 INFO [kafka-coordinator-heartbeat-thread | lxk_0514] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$HeartbeatResponseHandler.handle:855) - [Consumer clientId=consumer-1, groupId=lxk_0514] Attempt to heartbeat failed since group is rebalancing
+14 May 2020 12:02:59,720 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinPrepare:472) - [Consumer clientId=consumer-1, groupId=lxk_0514] Revoking previously assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 12:02:59,721 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsRevoked:785) - topic CONNECTION-RECORD-LOG - partition 0 revoked.
+14 May 2020 12:02:59,722 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator.sendJoinGroupRequest:509) - [Consumer clientId=consumer-1, groupId=lxk_0514] (Re-)joining group
+14 May 2020 12:02:59,729 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess:473) - [Consumer clientId=consumer-1, groupId=lxk_0514] Successfully joined group with generation 21
+14 May 2020 12:02:59,730 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete:280) - [Consumer clientId=consumer-1, groupId=lxk_0514] Setting newly assigned partitions [CONNECTION-RECORD-LOG-0]
+14 May 2020 12:02:59,730 INFO [SinkRunner-PollingRunner-DefaultSinkProcessor] (org.apache.flume.channel.kafka.ChannelRebalanceListener.onPartitionsAssigned:792) - topic CONNECTION-RECORD-LOG - partition 0 assigned.
diff --git a/external-flume/conf/security/pro/2f2k.properties b/external-flume/conf_external/traffic/pro/2f2k.properties
index 5e5c573..5e5c573 100644
--- a/external-flume/conf/security/pro/2f2k.properties
+++ b/external-flume/conf_external/traffic/pro/2f2k.properties
diff --git a/external-flume/conf/security/pro/4f2k.properties b/external-flume/conf_external/traffic/pro/4f2k.properties
index bf0408b..bf0408b 100644
--- a/external-flume/conf/security/pro/4f2k.properties
+++ b/external-flume/conf_external/traffic/pro/4f2k.properties
diff --git a/external-flume/conf/security/pro/f2k.properties b/external-flume/conf_external/traffic/pro/f2k.properties
index 7047e42..7047e42 100644
--- a/external-flume/conf/security/pro/f2k.properties
+++ b/external-flume/conf_external/traffic/pro/f2k.properties
diff --git a/external-flume/conf/security/pro/f2k1.properties b/external-flume/conf_external/traffic/pro/f2k1.properties
index 2dadced..2dadced 100644
--- a/external-flume/conf/security/pro/f2k1.properties
+++ b/external-flume/conf_external/traffic/pro/f2k1.properties
diff --git a/external-flume/conf/security/pro/f2k2.properties b/external-flume/conf_external/traffic/pro/f2k2.properties
index e17c3e2..e17c3e2 100644
--- a/external-flume/conf/security/pro/f2k2.properties
+++ b/external-flume/conf_external/traffic/pro/f2k2.properties
diff --git a/external-flume/conf/security/pro/f2k3.properties b/external-flume/conf_external/traffic/pro/f2k3.properties
index 2dadced..2dadced 100644
--- a/external-flume/conf/security/pro/f2k3.properties
+++ b/external-flume/conf_external/traffic/pro/f2k3.properties
diff --git a/external-flume/conf/security/pro/file.properties b/external-flume/conf_external/traffic/pro/file.properties
index f6813e7..f6813e7 100644
--- a/external-flume/conf/security/pro/file.properties
+++ b/external-flume/conf_external/traffic/pro/file.properties
diff --git a/external-flume/conf/security/pro/file1.properties b/external-flume/conf_external/traffic/pro/file1.properties
index 17dac7f..17dac7f 100644
--- a/external-flume/conf/security/pro/file1.properties
+++ b/external-flume/conf_external/traffic/pro/file1.properties
diff --git a/external-flume/conf/security/pro/file2.properties b/external-flume/conf_external/traffic/pro/file2.properties
index f6813e7..f6813e7 100644
--- a/external-flume/conf/security/pro/file2.properties
+++ b/external-flume/conf_external/traffic/pro/file2.properties
diff --git a/external-flume/conf/security/pro/k2f.properties b/external-flume/conf_external/traffic/pro/k2f.properties
index 35d7d68..35d7d68 100644
--- a/external-flume/conf/security/pro/k2f.properties
+++ b/external-flume/conf_external/traffic/pro/k2f.properties
diff --git a/external-flume/conf/security/pro/kafka.properties b/external-flume/conf_external/traffic/pro/kafka.properties
index da560b2..da560b2 100644
--- a/external-flume/conf/security/pro/kafka.properties
+++ b/external-flume/conf_external/traffic/pro/kafka.properties
diff --git a/external-flume/conf/security/pro/kafka1.properties b/external-flume/conf_external/traffic/pro/kafka1.properties
index ed30b16..ed30b16 100644
--- a/external-flume/conf/security/pro/kafka1.properties
+++ b/external-flume/conf_external/traffic/pro/kafka1.properties
diff --git a/external-flume/conf/security/pro/kafka2.properties b/external-flume/conf_external/traffic/pro/kafka2.properties
index 6040a05..6040a05 100644
--- a/external-flume/conf/security/pro/kafka2.properties
+++ b/external-flume/conf_external/traffic/pro/kafka2.properties
diff --git a/external-flume/conf/security/pro/log4j.properties b/external-flume/conf_external/traffic/pro/log4j.properties
index c948bff..c948bff 100644
--- a/external-flume/conf/security/pro/log4j.properties
+++ b/external-flume/conf_external/traffic/pro/log4j.properties
diff --git a/external-flume/conf/security/pro/tail.properties b/external-flume/conf_external/traffic/pro/tail.properties
index c677318..c677318 100644
--- a/external-flume/conf/security/pro/tail.properties
+++ b/external-flume/conf_external/traffic/pro/tail.properties
diff --git a/external-flume/conf/security/restart_log/restart_connection_k2f.log b/external-flume/conf_external/traffic/restart_log/restart_connection_k2f.log
index 281d596..1ad7998 100644
--- a/external-flume/conf/security/restart_log/restart_connection_k2f.log
+++ b/external-flume/conf_external/traffic/restart_log/restart_connection_k2f.log
@@ -66,3 +66,7 @@
20200513192014 ---> the connection_k2f_APP restart ---> 1
##############################################################
20200513192212 ---> the connection_k2f_APP restart ---> 1
+##############################################################
+20200514112452 ---> the connection_k2f_APP restart ---> 1
+##############################################################
+20200514114641 ---> the connection_k2f_APP restart ---> 1
diff --git a/external-flume/conf/security/restart_log/restart_f2k.log b/external-flume/conf_external/traffic/restart_log/restart_f2k.log
index 9d5fe99..9d5fe99 100644
--- a/external-flume/conf/security/restart_log/restart_f2k.log
+++ b/external-flume/conf_external/traffic/restart_log/restart_f2k.log
diff --git a/external-flume/conf/security/restart_log/restart_security_f2k.log b/external-flume/conf_external/traffic/restart_log/restart_security_f2k.log
index 690385d..690385d 100644
--- a/external-flume/conf/security/restart_log/restart_security_f2k.log
+++ b/external-flume/conf_external/traffic/restart_log/restart_security_f2k.log
diff --git a/external-flume/conf_external/traffic/start_flume.sh b/external-flume/conf_external/traffic/start_flume.sh
new file mode 100755
index 0000000..01dde8b
--- /dev/null
+++ b/external-flume/conf_external/traffic/start_flume.sh
@@ -0,0 +1,5 @@
+#!/bin/sh
+
+BASE_DIR=$(cd $(dirname $0); pwd)
+DAE_FILE=$(cd $(dirname $0); ls | grep dae*)
+nohup ${BASE_DIR}/$DAE_FILE >/dev/null 2>&1 &
diff --git a/external-flume/conf/security/stop_flume.sh b/external-flume/conf_external/traffic/stop_flume.sh
index e4417aa..e4417aa 100755
--- a/external-flume/conf/security/stop_flume.sh
+++ b/external-flume/conf_external/traffic/stop_flume.sh
diff --git a/external-flume/conf_external/traffic/traffic_k2f.properties b/external-flume/conf_external/traffic/traffic_k2f.properties
new file mode 100644
index 0000000..132eeb9
--- /dev/null
+++ b/external-flume/conf_external/traffic/traffic_k2f.properties
@@ -0,0 +1,22 @@
+#为source channel sink起名
+connection.channels = c1
+connection.sinks = k1
+#指定channel为kafka channel,省略source使得效率更高
+connection.channels.c1.type = org.apache.flume.channel.kafka.KafkaChannel
+connection.channels.c1.kafka.bootstrap.servers = 192.168.40.152:9092
+connection.channels.c1.kafka.topic = TRAFFIC-METRICS-LOG
+connection.channels.c1.kafka.consumer.group.id = lxk_0514
+connection.channels.c1.kafka.consumer.auto.offset.reset = latest
+connection.channels.c1.kafka.consumer.enable.auto.commit = true
+connection.channels.c1.kafka.consumer.fetch.max.wait.ms = 1000
+connection.channels.c1.kafka.consumer.fetch.min.bytes = 1048576
+connection.channels.c1.parseAsFlumeEvent = false
+
+#原始file roll sink,将数据写到本机磁盘
+connection.sinks.k1.type = file_roll
+connection.sinks.k1.channel = c1
+connection.sinks.k1.sink.pathManager = default
+connection.sinks.k1.sink.pathManager.extension = dat
+connection.sinks.k1.sink.pathManager.prefix = test-
+connection.sinks.k1.sink.rollInterval = 30
+connection.sinks.k1.sink.directory = /home/192.168.60.101/TRAFFIC-METRICS-LOG
diff --git a/external-flume/flume/file/flume.tar b/external-flume/flume/file/flume.tar
new file mode 100644
index 0000000..861cb21
--- /dev/null
+++ b/external-flume/flume/file/flume.tar
Binary files differ
diff --git a/external-flume/flume/tasks/load_flume.yml b/external-flume/flume/tasks/load_flume.yml
new file mode 100644
index 0000000..4dd4d2f
--- /dev/null
+++ b/external-flume/flume/tasks/load_flume.yml
@@ -0,0 +1,115 @@
+- name: stop and remove {{ image_container_name }} container
+ docker_container:
+ name: '{{ image_container_name }}'
+ state: absent
+
+- name: remove old {{ image_name }} image
+ docker_image:
+ name: '{{ image_name }}'
+ tag: '{{ image_tag_name }}'
+ state: absent
+
+- name: create {{ soft_home_path }}
+ file:
+ state: directory
+ path: '{{ soft_home_path }}'
+
+- name: create {{ volume_path }}
+ file:
+ state: directory
+ path: '{{ volume_path }}/tsg3.0-volumes/flume/{{ machine_host }}/CONNECTION-RECORD-COMPLETED-LOG'
+
+- name: create {{ volume_path }}
+ file:
+ state: directory
+ path: '{{ volume_path }}/tsg3.0-volumes/flume/{{ machine_host }}/PROXY-EVENT-COMPLETED-LOG'
+
+- name: create {{ volume_path }}
+ file:
+ state: directory
+ path: '{{ volume_path }}/tsg3.0-volumes/flume/{{ machine_host }}/ACTIVE-DEFENCE-EVENT-COMPLETED-LOG'
+
+- name: create {{ volume_path }}
+ file:
+ state: directory
+ path: '{{ volume_path }}/tsg3.0-volumes/flume/{{ machine_host }}/SECURITY-EVENT-COMPLETED-LOG'
+
+- name: create {{ volume_path }}
+ file:
+ state: directory
+ path: '{{ volume_path }}/tsg3.0-volumes/flume/{{ machine_host }}/TRAFFIC-METRICS-LOG'
+
+- name: copy {{ image_tar_name }}
+ copy:
+ src: '{{ images_path }}/{{ image_tar_name }}'
+ dest: '{{ soft_home_path }}/{{ image_tar_name }}'
+ force: true
+ backup: yes
+ when: offline_install #离线模式从本地load文件加载
+
+- name: copy flume conf files
+ copy:
+ src: 'file/flume.tar'
+ dest: '{{ volume_path }}/tsg3.0-volumes/flume/flume.tar'
+ force: true
+ backup: yes
+
+- name: tar -zxvf flume.tar
+ shell: cd {{ volume_path }}/tsg3.0-volumes/flume/ && tar -zxvf flume.tar && rm -rf flume.tar
+
+- name: load new {{ image_name }} image from {{ image_tar_name }}
+ docker_image:
+ name: '{{ image_name }}'
+ tag: '{{ image_tag_name }}'
+ load_path: '{{ soft_home_path }}/{{ image_tar_name }}'
+ source: load
+ force_tag: yes
+ force_source: yes
+ when: offline_install #离线模式从本地load文件加载
+
+- name: load new {{ image_name }} image from docker-registry
+ docker_image:
+ name: '{{ docker_registry_image_and_tag }}'
+ source: pull
+ force_tag: yes
+ force_source: yes
+ when: offline_install==false #在线模式从docker-registry中pull镜像
+
+- name: copy {{ image_name }} docker-compose.yml
+ template:
+ src: docker-compose.yml.j2
+ dest: '{{ soft_home_path }}/docker-compose.yml'
+ backup: yes
+
+- name: copy k2fConnRecord.properties.j2
+ template:
+ src: k2fConnRecord.properties.j2
+ dest: '{{ volume_path }}/tsg3.0-volumes/flume/conf/connRecordConf/k2fConnRecord.properties'
+ backup: yes
+
+- name: copy k2fProxyEvent.properties.j2
+ template:
+ src: k2fProxyEvent.properties.j2
+ dest: '{{ volume_path }}/tsg3.0-volumes/flume/conf/proxyEventConf/k2fProxyEvent.properties'
+ backup: yes
+
+- name: copy k2fTraffic.properties.j2
+ template:
+ src: k2fTraffic.properties.j2
+ dest: '{{ volume_path }}/tsg3.0-volumes/flume/conf/trafficMetricsConf/k2fTraffic.properties'
+ backup: yes
+
+- name: copy k2fSecEvent.properties.j2
+ template:
+ src: k2fSecEvent.properties.j2
+ dest: '{{ volume_path }}/tsg3.0-volumes/flume/conf/securityEventConf/k2fSecEvent.properties'
+ backup: yes
+
+- name: copy k2fActDef.properties.j2
+ template:
+ src: k2fActDef.properties.j2
+ dest: '{{ volume_path }}/tsg3.0-volumes/flume/conf/actDefConf/k2fActDef.properties'
+ backup: yes
+- name: start {{ image_container_name }} container
+ docker_compose:
+ project_src: '{{ soft_home_path }}'
diff --git a/external-flume/flume/tasks/main.yml b/external-flume/flume/tasks/main.yml
new file mode 100644
index 0000000..886a51f
--- /dev/null
+++ b/external-flume/flume/tasks/main.yml
@@ -0,0 +1 @@
+- include: load_flume.yml
diff --git a/external-flume/flume/templates/docker-compose.yml.j2 b/external-flume/flume/templates/docker-compose.yml.j2
new file mode 100644
index 0000000..12beb8a
--- /dev/null
+++ b/external-flume/flume/templates/docker-compose.yml.j2
@@ -0,0 +1,24 @@
+version: '2'
+
+services:
+ flume:
+{% if offline_install %}
+ image: {{ image_name }}:{{ image_tag_name }}
+{% else %}
+ image: {{ docker_registry_image_and_tag }}
+{% endif %}
+ restart: always
+ container_name: {{ image_container_name }}
+ environment:
+ SESPROCESSNUM: 1
+ SESPORTBEGIN: 35000
+ volumes:
+ - "{{ volume_path }}/tsg3.0-volumes/flume/conf:/home/apache-flume-1.9.0-bin/conf"
+ - "{{ volume_path }}/tsg3.0-volumes/flume/flumeController:/home/apache-flume-1.9.0-bin/flumeController"
+ - "{{ volume_path }}/tsg3.0-volumes/flume/flumeMonitor:/home/apache-flume-1.9.0-bin/flumeMonitor"
+ - "{{ volume_path }}/tsg3.0-volumes/flume/plugins.d:/home/apache-flume-1.9.0-bin/plugins.d"
+ - "{{ volume_path }}/tsg3.0-volumes/flume/dat/:/home/dat/"
+ - "/etc/localtime:/etc/localtime:ro"
+ - "/etc/timezone:/etc/timezone:ro"
+ network_mode: "host"
+
diff --git a/external-flume/flume/templates/k2fActDef.properties.j2 b/external-flume/flume/templates/k2fActDef.properties.j2
new file mode 100644
index 0000000..7bb3333
--- /dev/null
+++ b/external-flume/flume/templates/k2fActDef.properties.j2
@@ -0,0 +1,75 @@
+#为source channel sink起名
+plactdef.sources = kafkasource1
+plactdef.channels = channel1
+plactdef.sinks = kafkasink1
+#指定source收集到的数据发送到哪个管道
+plactdef.sources.kafkasource1.channels = channel1
+#指定source数据收集策略
+plactdef.sources.kafkasource1.type = org.apache.flume.source.kafka.KafkaSource
+plactdef.sources.kafkasource1.batchSize = 30000
+plactdef.sources.kafkasource1.batchDurationMillis = 1000
+plactdef.sources.kafkasource1.kafka.bootstrap.servers = {{ machine_host }}:9092
+plactdef.sources.kafkasource1.kafka.topics = ACTIVE-DEFENCE-EVENT-LOG
+plactdef.sources.kafkasource1.kafka.consumer.group.id = testactin_1
+
+#指定channel为memory,即表示所有的数据都装进memory当中
+plactdef.channels.channel1.type = memory
+plactdef.channels.channel1.capacity = 2000000
+plactdef.channels.channel1.transactionCapacity = 30000
+plactdef.channels.channel1.byteCapacityBufferPercentage = 40
+plactdef.channels.channel1.byteCapacity = 2147483648
+#原始file roll sink
+plactdef.sinks.kafkasink1.type = file_roll
+plactdef.sinks.kafkasink1.channel = channel1
+plactdef.sinks.kafkasink1.sink.pathManager = default
+plactdef.sinks.kafkasink1.sink.pathManager.extension = dat
+plactdef.sinks.kafkasink1.sink.pathManager.prefix = active-
+plactdef.sinks.kafkasink1.sink.rollInterval = 10
+plactdef.sinks.kafkasink1.sink.directory = /home/{{ machine_host }}/ACTIVE-DEFENCE-EVENT-COMPLETED-LOG
+
+#拦截器批量设置
+#zx
+#plactdef.sources.kafkasource1.interceptors = i1 i4
+#plactdef.sources.kafkasource1.interceptors = i5 i6
+plactdef.sources.kafkasource1.interceptors = i6
+#拦截器-1-定位库
+plactdef.sources.kafkasource1.interceptors.i1.type = cn.ac.iie.flume.interceptor.FlumeIPAsnApp$FlumeIpAsnAppBuilder
+plactdef.sources.kafkasource1.interceptors.i1.ipDat = /home/dat/Kazakhstan.mmdb
+plactdef.sources.kafkasource1.interceptors.i1.v4Asn = /home/dat/asn_v4.mmdb
+plactdef.sources.kafkasource1.interceptors.i1.v6Asn = /home/dat/asn_v6.mmdb
+plactdef.sources.kafkasource1.interceptors.i1.serveripField = common_server_ip
+plactdef.sources.kafkasource1.interceptors.i1.clientipField = common_client_ip
+plactdef.sources.kafkasource1.interceptors.i1.serverlocatField = common_server_location
+plactdef.sources.kafkasource1.interceptors.i1.clientlocatField = common_client_location
+plactdef.sources.kafkasource1.interceptors.i1.serverasnField = common_server_asn
+plactdef.sources.kafkasource1.interceptors.i1.clientasnField = common_client_asn
+#plactdef.sources.kafkasource1.interceptors.i1.customipField = nocusipfield
+#plactdef.sources.kafkasource1.interceptors.i1.customlocatField = nocuslocatfield
+#plactdef.sources.kafkasource1.interceptors.i1.customasnField = nocusasnfield
+#拦截器-2-域名
+plactdef.sources.kafkasource1.interceptors.i2.type = cn.ac.iie.flume.interceptor.FlumeDomainApp$FlumeDomainAppBuilder
+plactdef.sources.kafkasource1.interceptors.i2.domainField = domain
+#拦截器-3-分布式ID与时间戳
+plactdef.sources.kafkasource1.interceptors.i3.type = cn.ac.iie.flume.interceptor.FlumeUidApp$FlumeUidAppBuilder
+plactdef.sources.kafkasource1.interceptors.i3.uId = uid
+plactdef.sources.kafkasource1.interceptors.i3.zookeeperIp = {{ machine_host }}:2181
+plactdef.sources.kafkasource1.interceptors.i3.kafkaTopic = SESSION-TEST-LOG
+plactdef.sources.kafkasource1.interceptors.i3.timeStamp = recv_time
+plactdef.sources.kafkasource1.interceptors.i3.dataCenterIdNum = 15
+#拦截器-4-app_label截取映射
+plactdef.sources.kafkasource1.interceptors.i4.type = cn.ac.iie.flume.interceptor.FlumeFieldSplitterApp$FlumeFieldSplitterAppBuilder
+plactdef.sources.kafkasource1.interceptors.i4.handleField = app_label
+plactdef.sources.kafkasource1.interceptors.i4.splitFields = proto:PROTO_ID;app:APP_ID;web:WEB_ID
+plactdef.sources.kafkasource1.interceptors.i4.fieldsSeparator = ;
+plactdef.sources.kafkasource1.interceptors.i4.fieldDelimiter = =
+#拦截器-5-avro反序列化
+plactdef.sources.kafkasource1.interceptors.i5.type = com.zdjizhi.flume.interceptor.FlumeAvroDeserialiApp$FlumeAvroDeserialiAppBuilder
+plactdef.sources.kafkasource1.interceptors.i5.avroSchemaUrl = http://{{ machine_host }}:9999/metadata/schema/v1/fields/active_defence_event_log
+#拦截器-6-动态补全
+plactdef.sources.kafkasource1.interceptors.i6.type = com.zdjizhi.flume.interceptor.FlumeDynamicApp$FlumeDynamicAppBuilder
+plactdef.sources.kafkasource1.interceptors.i6.schemaHttpUrl = http://{{ machine_host }}:9999/metadata/schema/v1/fields/active_defence_event_log
+plactdef.sources.kafkasource1.interceptors.i6.uidZookeeperIp = {{ machine_host }}:2181
+plactdef.sources.kafkasource1.interceptors.i6.dataCenterIdNum = 5
+plactdef.sources.kafkasource1.interceptors.i6.ipDatPath = /home/dat
+plactdef.sources.kafkasource1.interceptors.i6.hbaseZookeeperIp = {{ machine_host }}:2181
+plactdef.sources.kafkasource1.interceptors.i6.hbaseTableName = subscriber_info
diff --git a/external-flume/flume/templates/k2fConnRecord.properties.j2 b/external-flume/flume/templates/k2fConnRecord.properties.j2
new file mode 100644
index 0000000..9aa0a19
--- /dev/null
+++ b/external-flume/flume/templates/k2fConnRecord.properties.j2
@@ -0,0 +1,76 @@
+#为source channel sink起名
+plconnrecord.sources = kafkasource1
+plconnrecord.channels = channel1
+plconnrecord.sinks = kafkasink1
+#指定source收集到的数据发送到哪个管道
+plconnrecord.sources.kafkasource1.channels = channel1
+#指定source数据收集策略
+plconnrecord.sources.kafkasource1.type = org.apache.flume.source.kafka.KafkaSource
+plconnrecord.sources.kafkasource1.batchSize = 30000
+plconnrecord.sources.kafkasource1.batchDurationMillis = 1000
+plconnrecord.sources.kafkasource1.kafka.bootstrap.servers = {{ machine_host }}:9092
+plconnrecord.sources.kafkasource1.kafka.topics = CONNECTION-RECORD-LOG
+plconnrecord.sources.kafkasource1.kafka.consumer.group.id = testconnin_1
+
+#指定channel为memory,即表示所有的数据都装进memory当中
+plconnrecord.channels.channel1.type = memory
+plconnrecord.channels.channel1.capacity = 2000000
+plconnrecord.channels.channel1.transactionCapacity = 30000
+plconnrecord.channels.channel1.byteCapacityBufferPercentage = 40
+plconnrecord.channels.channel1.byteCapacity = 2147483648
+
+#指定sink为file roll sink
+plconnrecord.sinks.kafkasink1.type = file_roll
+plconnrecord.sinks.kafkasink1.channel = channel1
+plconnrecord.sinks.kafkasink1.sink.pathManager = default
+plconnrecord.sinks.kafkasink1.sink.pathManager.extension = dat
+plconnrecord.sinks.kafkasink1.sink.pathManager.prefix = connection-
+plconnrecord.sinks.kafkasink1.sink.rollInterval = 10
+plconnrecord.sinks.kafkasink1.sink.directory = /home/{{ machine_host }}/CONNECTION-RECORD-COMPLETED-LOG
+
+#拦截器批量设置
+#zx
+#plconnrecord.sources.kafkasource1.interceptors = i1 i4
+#plconnrecord.sources.kafkasource1.interceptors = i5 i6
+plconnrecord.sources.kafkasource1.interceptors = i6
+#拦截器-1-定位库
+plconnrecord.sources.kafkasource1.interceptors.i1.type = cn.ac.iie.flume.interceptor.FlumeIPAsnApp$FlumeIpAsnAppBuilder
+plconnrecord.sources.kafkasource1.interceptors.i1.ipDat = /home/dat/Kazakhstan.mmdb
+plconnrecord.sources.kafkasource1.interceptors.i1.v4Asn = /home/dat/asn_v4.mmdb
+plconnrecord.sources.kafkasource1.interceptors.i1.v6Asn = /home/dat/asn_v6.mmdb
+plconnrecord.sources.kafkasource1.interceptors.i1.serveripField = common_server_ip
+plconnrecord.sources.kafkasource1.interceptors.i1.clientipField = common_client_ip
+plconnrecord.sources.kafkasource1.interceptors.i1.serverlocatField = common_server_location
+plconnrecord.sources.kafkasource1.interceptors.i1.clientlocatField = common_client_location
+plconnrecord.sources.kafkasource1.interceptors.i1.serverasnField = common_server_asn
+plconnrecord.sources.kafkasource1.interceptors.i1.clientasnField = common_client_asn
+#plconnrecord.sources.kafkasource1.interceptors.i1.customipField = nocusipfield
+#plconnrecord.sources.kafkasource1.interceptors.i1.customlocatField = nocuslocatfield
+#plconnrecord.sources.kafkasource1.interceptors.i1.customasnField = nocusasnfield
+#拦截器-2-域名
+plconnrecord.sources.kafkasource1.interceptors.i2.type = cn.ac.iie.flume.interceptor.FlumeDomainApp$FlumeDomainAppBuilder
+plconnrecord.sources.kafkasource1.interceptors.i2.domainField = domain
+#拦截器-3-分布式ID与时间戳
+plconnrecord.sources.kafkasource1.interceptors.i3.type = cn.ac.iie.flume.interceptor.FlumeUidApp$FlumeUidAppBuilder
+plconnrecord.sources.kafkasource1.interceptors.i3.uId = uid
+plconnrecord.sources.kafkasource1.interceptors.i3.zookeeperIp = {{ machine_host }}:2181
+plconnrecord.sources.kafkasource1.interceptors.i3.kafkaTopic = SESSION-TEST-LOG
+plconnrecord.sources.kafkasource1.interceptors.i3.timeStamp = recv_time
+plconnrecord.sources.kafkasource1.interceptors.i3.dataCenterIdNum = 15
+#拦截器-4-app_label截取映射
+plconnrecord.sources.kafkasource1.interceptors.i4.type = cn.ac.iie.flume.interceptor.FlumeFieldSplitterApp$FlumeFieldSplitterAppBuilder
+plconnrecord.sources.kafkasource1.interceptors.i4.handleField = app_label
+plconnrecord.sources.kafkasource1.interceptors.i4.splitFields = proto:PROTO_ID;app:APP_ID;web:WEB_ID
+plconnrecord.sources.kafkasource1.interceptors.i4.fieldsSeparator = ;
+plconnrecord.sources.kafkasource1.interceptors.i4.fieldDelimiter = =
+#拦截器-5-avro反序列化
+plconnrecord.sources.kafkasource1.interceptors.i5.type = com.zdjizhi.flume.interceptor.FlumeAvroDeserialiApp$FlumeAvroDeserialiAppBuilder
+plconnrecord.sources.kafkasource1.interceptors.i5.avroSchemaUrl = http://{{ machine_host }}:9999/metadata/schema/v1/fields/connection_record_log
+#拦截器-6-动态补全
+plconnrecord.sources.kafkasource1.interceptors.i6.type = com.zdjizhi.flume.interceptor.FlumeDynamicApp$FlumeDynamicAppBuilder
+plconnrecord.sources.kafkasource1.interceptors.i6.schemaHttpUrl = http://{{ machine_host }}:9999/metadata/schema/v1/fields/connection_record_log
+plconnrecord.sources.kafkasource1.interceptors.i6.uidZookeeperIp = {{ machine_host }}:2181
+plconnrecord.sources.kafkasource1.interceptors.i6.dataCenterIdNum = 0
+plconnrecord.sources.kafkasource1.interceptors.i6.ipDatPath = /home/dat
+plconnrecord.sources.kafkasource1.interceptors.i6.hbaseZookeeperIp = {{ machine_host }}:2181
+plconnrecord.sources.kafkasource1.interceptors.i6.hbaseTableName = subscriber_info
diff --git a/external-flume/flume/templates/k2fProxyEvent.properties.j2 b/external-flume/flume/templates/k2fProxyEvent.properties.j2
new file mode 100644
index 0000000..7983ec2
--- /dev/null
+++ b/external-flume/flume/templates/k2fProxyEvent.properties.j2
@@ -0,0 +1,75 @@
+#为source channel sink起名
+plproxyevent.sources = kafkasource1
+plproxyevent.channels = channel1
+plproxyevent.sinks = kafkasink1
+#指定source收集到的数据发送到哪个管道
+plproxyevent.sources.kafkasource1.channels = channel1
+#指定source数据收集策略
+plproxyevent.sources.kafkasource1.type = org.apache.flume.source.kafka.KafkaSource
+plproxyevent.sources.kafkasource1.batchSize = 30000
+plproxyevent.sources.kafkasource1.batchDurationMillis = 1000
+plproxyevent.sources.kafkasource1.kafka.bootstrap.servers = {{ machine_host }}:9092
+plproxyevent.sources.kafkasource1.kafka.topics = PROXY-EVENT-LOG
+plproxyevent.sources.kafkasource1.kafka.consumer.group.id = testproxyin_1
+
+#指定channel为memory,即表示所有的数据都装进memory当中
+plproxyevent.channels.channel1.type = memory
+plproxyevent.channels.channel1.capacity = 2000000
+plproxyevent.channels.channel1.transactionCapacity = 30000
+plproxyevent.channels.channel1.byteCapacityBufferPercentage = 40
+plproxyevent.channels.channel1.byteCapacity = 2147483648
+#原始file roll sink
+plproxyevent.sinks.kafkasink1.type = file_roll
+plproxyevent.sinks.kafkasink1.channel = channel1
+plproxyevent.sinks.kafkasink1.sink.pathManager = default
+plproxyevent.sinks.kafkasink1.sink.pathManager.extension = dat
+plproxyevent.sinks.kafkasink1.sink.pathManager.prefix = proxy-
+plproxyevent.sinks.kafkasink1.sink.rollInterval = 10
+plproxyevent.sinks.kafkasink1.sink.directory = /home/{{ machine_host }}/PROXY-EVENT-COMPLETED-LOG
+
+#拦截器批量设置
+#zx
+#plproxyevent.sources.kafkasource1.interceptors = i1 i4
+#plproxyevent.sources.kafkasource1.interceptors = i5 i6
+plproxyevent.sources.kafkasource1.interceptors = i6
+#拦截器-1-定位库
+plproxyevent.sources.kafkasource1.interceptors.i1.type = cn.ac.iie.flume.interceptor.FlumeIPAsnApp$FlumeIpAsnAppBuilder
+plproxyevent.sources.kafkasource1.interceptors.i1.ipDat = /home/dat/Kazakhstan.mmdb
+plproxyevent.sources.kafkasource1.interceptors.i1.v4Asn = /home/dat/asn_v4.mmdb
+plproxyevent.sources.kafkasource1.interceptors.i1.v6Asn = /home/dat/asn_v6.mmdb
+plproxyevent.sources.kafkasource1.interceptors.i1.serveripField = common_server_ip
+plproxyevent.sources.kafkasource1.interceptors.i1.clientipField = common_client_ip
+plproxyevent.sources.kafkasource1.interceptors.i1.serverlocatField = common_server_location
+plproxyevent.sources.kafkasource1.interceptors.i1.clientlocatField = common_client_location
+plproxyevent.sources.kafkasource1.interceptors.i1.serverasnField = common_server_asn
+plproxyevent.sources.kafkasource1.interceptors.i1.clientasnField = common_client_asn
+#plproxyevent.sources.kafkasource1.interceptors.i1.customipField = nocusipfield
+#plproxyevent.sources.kafkasource1.interceptors.i1.customlocatField = nocuslocatfield
+#plproxyevent.sources.kafkasource1.interceptors.i1.customasnField = nocusasnfield
+#拦截器-2-域名
+plproxyevent.sources.kafkasource1.interceptors.i2.type = cn.ac.iie.flume.interceptor.FlumeDomainApp$FlumeDomainAppBuilder
+plproxyevent.sources.kafkasource1.interceptors.i2.domainField = domain
+#拦截器-3-分布式ID与时间戳
+plproxyevent.sources.kafkasource1.interceptors.i3.type = cn.ac.iie.flume.interceptor.FlumeUidApp$FlumeUidAppBuilder
+plproxyevent.sources.kafkasource1.interceptors.i3.uId = uid
+plproxyevent.sources.kafkasource1.interceptors.i3.zookeeperIp = {{ machine_host }}:2181
+plproxyevent.sources.kafkasource1.interceptors.i3.kafkaTopic = SESSION-TEST-LOG
+plproxyevent.sources.kafkasource1.interceptors.i3.timeStamp = recv_time
+plproxyevent.sources.kafkasource1.interceptors.i3.dataCenterIdNum = 15
+#拦截器-4-app_label截取映射
+plproxyevent.sources.kafkasource1.interceptors.i4.type = cn.ac.iie.flume.interceptor.FlumeFieldSplitterApp$FlumeFieldSplitterAppBuilder
+plproxyevent.sources.kafkasource1.interceptors.i4.handleField = app_label
+plproxyevent.sources.kafkasource1.interceptors.i4.splitFields = proto:PROTO_ID;app:APP_ID;web:WEB_ID
+plproxyevent.sources.kafkasource1.interceptors.i4.fieldsSeparator = ;
+plproxyevent.sources.kafkasource1.interceptors.i4.fieldDelimiter = =
+#拦截器-5-avro反序列化
+plproxyevent.sources.kafkasource1.interceptors.i5.type = com.zdjizhi.flume.interceptor.FlumeAvroDeserialiApp$FlumeAvroDeserialiAppBuilder
+plproxyevent.sources.kafkasource1.interceptors.i5.avroSchemaUrl = http://{{ machine_host }}:9999/metadata/schema/v1/fields/proxy_event_log
+#拦截器-6-动态补全
+plproxyevent.sources.kafkasource1.interceptors.i6.type = com.zdjizhi.flume.interceptor.FlumeDynamicApp$FlumeDynamicAppBuilder
+plproxyevent.sources.kafkasource1.interceptors.i6.schemaHttpUrl = http://{{ machine_host }}:9999/metadata/schema/v1/fields/proxy_event_log
+plproxyevent.sources.kafkasource1.interceptors.i6.uidZookeeperIp = {{ machine_host }}:2181
+plproxyevent.sources.kafkasource1.interceptors.i6.dataCenterIdNum = 1
+plproxyevent.sources.kafkasource1.interceptors.i6.ipDatPath = /home/dat
+plproxyevent.sources.kafkasource1.interceptors.i6.hbaseZookeeperIp = {{ machine_host }}:2181
+plproxyevent.sources.kafkasource1.interceptors.i6.hbaseTableName = subscriber_info
diff --git a/external-flume/flume/templates/k2fSecEvent.properties.j2 b/external-flume/flume/templates/k2fSecEvent.properties.j2
new file mode 100644
index 0000000..0c18492
--- /dev/null
+++ b/external-flume/flume/templates/k2fSecEvent.properties.j2
@@ -0,0 +1,75 @@
+#为source channel sink起名
+plsecevent.sources = kafkasource1
+plsecevent.channels = channel1
+plsecevent.sinks = kafkasink1
+#指定source收集到的数据发送到哪个管道
+plsecevent.sources.kafkasource1.channels = channel1
+#指定source数据收集策略
+plsecevent.sources.kafkasource1.type = org.apache.flume.source.kafka.KafkaSource
+plsecevent.sources.kafkasource1.batchSize = 30000
+plsecevent.sources.kafkasource1.batchDurationMillis = 1000
+plsecevent.sources.kafkasource1.kafka.bootstrap.servers = {{ machine_host }}:9092
+plsecevent.sources.kafkasource1.kafka.topics = SECURITY-EVENT-LOG
+plsecevent.sources.kafkasource1.kafka.consumer.group.id = testsecin_1
+
+#指定channel为memory,即表示所有的数据都装进memory当中
+plsecevent.channels.channel1.type = memory
+plsecevent.channels.channel1.capacity = 2000000
+plsecevent.channels.channel1.transactionCapacity = 30000
+plsecevent.channels.channel1.byteCapacityBufferPercentage = 40
+plsecevent.channels.channel1.byteCapacity = 2147483648
+#原始file roll sink
+plsecevent.sinks.kafkasink1.type = file_roll
+plsecevent.sinks.kafkasink1.channel = channel1
+plsecevent.sinks.kafkasink1.sink.pathManager = default
+plsecevent.sinks.kafkasink1.sink.pathManager.extension = dat
+plsecevent.sinks.kafkasink1.sink.pathManager.prefix = security-
+plsecevent.sinks.kafkasink1.sink.rollInterval = 10
+plsecevent.sinks.kafkasink1.sink.directory = /home/{{ machine_host }}/SECURITY-EVENT-COMPLETED-LOG
+
+#拦截器批量设置
+#zx
+#plsecevent.sources.kafkasource1.interceptors = i1 i4
+#plsecevent.sources.kafkasource1.interceptors = i5 i6
+plsecevent.sources.kafkasource1.interceptors = i6
+#拦截器-1-定位库
+plsecevent.sources.kafkasource1.interceptors.i1.type = cn.ac.iie.flume.interceptor.FlumeIPAsnApp$FlumeIpAsnAppBuilder
+plsecevent.sources.kafkasource1.interceptors.i1.ipDat = /home/dat/Kazakhstan.mmdb
+plsecevent.sources.kafkasource1.interceptors.i1.v4Asn = /home/dat/asn_v4.mmdb
+plsecevent.sources.kafkasource1.interceptors.i1.v6Asn = /home/dat/asn_v6.mmdb
+plsecevent.sources.kafkasource1.interceptors.i1.serveripField = common_server_ip
+plsecevent.sources.kafkasource1.interceptors.i1.clientipField = common_client_ip
+plsecevent.sources.kafkasource1.interceptors.i1.serverlocatField = common_server_location
+plsecevent.sources.kafkasource1.interceptors.i1.clientlocatField = common_client_location
+plsecevent.sources.kafkasource1.interceptors.i1.serverasnField = common_server_asn
+plsecevent.sources.kafkasource1.interceptors.i1.clientasnField = common_client_asn
+#plsecevent.sources.kafkasource1.interceptors.i1.customipField = nocusipfield
+#plsecevent.sources.kafkasource1.interceptors.i1.customlocatField = nocuslocatfield
+#plsecevent.sources.kafkasource1.interceptors.i1.customasnField = nocusasnfield
+#拦截器-2-域名
+plsecevent.sources.kafkasource1.interceptors.i2.type = cn.ac.iie.flume.interceptor.FlumeDomainApp$FlumeDomainAppBuilder
+plsecevent.sources.kafkasource1.interceptors.i2.domainField = domain
+#拦截器-3-分布式ID与时间戳
+plsecevent.sources.kafkasource1.interceptors.i3.type = cn.ac.iie.flume.interceptor.FlumeUidApp$FlumeUidAppBuilder
+plsecevent.sources.kafkasource1.interceptors.i3.uId = uid
+plsecevent.sources.kafkasource1.interceptors.i3.zookeeperIp = {{ machine_host }}:2181
+plsecevent.sources.kafkasource1.interceptors.i3.kafkaTopic = SESSION-TEST-LOG
+plsecevent.sources.kafkasource1.interceptors.i3.timeStamp = recv_time
+plsecevent.sources.kafkasource1.interceptors.i3.dataCenterIdNum = 15
+#拦截器-4-app_label截取映射
+plsecevent.sources.kafkasource1.interceptors.i4.type = cn.ac.iie.flume.interceptor.FlumeFieldSplitterApp$FlumeFieldSplitterAppBuilder
+plsecevent.sources.kafkasource1.interceptors.i4.handleField = app_label
+plsecevent.sources.kafkasource1.interceptors.i4.splitFields = proto:PROTO_ID;app:APP_ID;web:WEB_ID
+plsecevent.sources.kafkasource1.interceptors.i4.fieldsSeparator = ;
+plsecevent.sources.kafkasource1.interceptors.i4.fieldDelimiter = =
+#拦截器-5-avro反序列化
+plsecevent.sources.kafkasource1.interceptors.i5.type = com.zdjizhi.flume.interceptor.FlumeAvroDeserialiApp$FlumeAvroDeserialiAppBuilder
+plsecevent.sources.kafkasource1.interceptors.i5.avroSchemaUrl = http://{{ machine_host }}:9999/metadata/schema/v1/fields/security_event_log
+#拦截器-6-动态补全
+plsecevent.sources.kafkasource1.interceptors.i6.type = com.zdjizhi.flume.interceptor.FlumeDynamicApp$FlumeDynamicAppBuilder
+plsecevent.sources.kafkasource1.interceptors.i6.schemaHttpUrl = http://{{ machine_host }}:9999/metadata/schema/v1/fields/security_event_log
+plsecevent.sources.kafkasource1.interceptors.i6.uidZookeeperIp = {{ machine_host }}:2181
+plsecevent.sources.kafkasource1.interceptors.i6.dataCenterIdNum = 3
+plsecevent.sources.kafkasource1.interceptors.i6.ipDatPath = /home/dat
+plsecevent.sources.kafkasource1.interceptors.i6.hbaseZookeeperIp = {{ machine_host }}:2181
+plsecevent.sources.kafkasource1.interceptors.i6.hbaseTableName = subscriber_info
diff --git a/external-flume/flume/templates/k2fTraffic.properties.j2 b/external-flume/flume/templates/k2fTraffic.properties.j2
new file mode 100644
index 0000000..3e6f94f
--- /dev/null
+++ b/external-flume/flume/templates/k2fTraffic.properties.j2
@@ -0,0 +1,22 @@
+#为source channel sink起名
+traffic.channels = c1
+traffic.sinks = k1
+#指定channel为kafka channel,省略source使得效率更高
+traffic.channels.c1.type = org.apache.flume.channel.kafka.KafkaChannel
+traffic.channels.c1.kafka.bootstrap.servers = {{ machine_host }}:9092
+traffic.channels.c1.kafka.topic = TRAFFIC-METRICS-LOG
+traffic.channels.c1.kafka.consumer.group.id = lxk_0514
+traffic.channels.c1.kafka.consumer.auto.offset.reset = latest
+traffic.channels.c1.kafka.consumer.enable.auto.commit = true
+traffic.channels.c1.kafka.consumer.fetch.max.wait.ms = 1000
+traffic.channels.c1.kafka.consumer.fetch.min.bytes = 1048576
+traffic.channels.c1.parseAsFlumeEvent = false
+
+#原始file roll sink,将数据写到本机磁盘
+traffic.sinks.k1.type = file_roll
+traffic.sinks.k1.channel = c1
+traffic.sinks.k1.sink.pathManager = default
+traffic.sinks.k1.sink.pathManager.extension = dat
+traffic.sinks.k1.sink.pathManager.prefix = traffic-
+traffic.sinks.k1.sink.rollInterval = 10
+traffic.sinks.k1.sink.directory = /home/{{ machine_host }}/TRAFFIC-METRICS-LOG
diff --git a/external-flume/flume/vars/main.yml b/external-flume/flume/vars/main.yml
new file mode 100644
index 0000000..d11e59c
--- /dev/null
+++ b/external-flume/flume/vars/main.yml
@@ -0,0 +1,25 @@
+#注意这里面定义的key中不能包含减号-,否则程序无法识别定义的key
+
+#########################从docker仓库pull镜像的相关属性##########################################################
+#docker仓库(harbor)中的地址,镜像名称和tag
+docker_registry_image_and_tag: 192.168.40.153:9080/tsg/flume:1.9.0 #bifang_api镜像和tag名称
+
+#########################离线加载docker镜像的相关属性##########################################################
+#离线模式下使用:bifang-api的docker镜像文件名称,在../file中的文件名称
+image_tar_name: flume-1.9.0.tar
+
+
+
+##########################公共属性#############################
+#设置容器的hostname
+#container_hostname: bifang-api
+#{{ image_tar_name }}中的镜像名称,或者{{ docker_registry_image_and_tag }}中的镜像名称,不含版本号
+image_name: flume
+#{{ image_tar_name }}的镜像版本号,或者{{ docker_registry_image_and_tag }}中的镜像版本号
+image_tag_name: 1.9.0
+#{{ image_tar_name }}的镜像,或者{{ docker_registry_image_and_tag }}中的镜像启动后的容器名称
+image_container_name: flume
+#镜像文件,docker-compose文件以及数据卷共享目录所在的位置,注意路径最后面不带/
+soft_home_path: /home/test/flume
+dll_volume_path: {ACTIVE-DEFENCE-EVENT-COMPLETED-LOG,CONNECTION-RECORD-COMPLETED-LOG,PROXY-EVENT-COMPLETED-LOG,SECURITY-EVENT-COMPLETED-LOG,TRAFFIC-METRICS-LOG}
+
diff --git a/internal-flume/ansible_role/flume/file/flume.tar b/internal-flume/ansible_role/flume/file/flume.tar
new file mode 100644
index 0000000..6ed06cd
--- /dev/null
+++ b/internal-flume/ansible_role/flume/file/flume.tar
Binary files differ
diff --git a/internal-flume/ansible_role/flume/tasks/load_flume.yml b/internal-flume/ansible_role/flume/tasks/load_flume.yml
new file mode 100644
index 0000000..1a794f9
--- /dev/null
+++ b/internal-flume/ansible_role/flume/tasks/load_flume.yml
@@ -0,0 +1,66 @@
+- name: stop and remove {{ image_container_name }} container
+ docker_container:
+ name: '{{ image_container_name }}'
+ state: absent
+
+- name: remove old {{ image_name }} image
+ docker_image:
+ name: '{{ image_name }}'
+ tag: '{{ image_tag_name }}'
+ state: absent
+
+- name: create {{ soft_home_path }}
+ file:
+ state: directory
+ path: '{{ soft_home_path }}'
+
+- name: create {{ volume_path }}
+ file:
+ state: directory
+ path: '{{ volume_path }}/tsg3.0-volumes/flume'
+
+- name: copy {{ image_tar_name }}
+ copy:
+ src: '{{ images_path }}/{{ image_tar_name }}'
+ dest: '{{ soft_home_path }}/{{ image_tar_name }}'
+ force: true
+ backup: yes
+ when: offline_install #离线模式从本地load文件加载
+
+- name: copy flume conf files
+ copy:
+ src: 'file/flume.tar'
+ dest: '{{ volume_path }}/tsg3.0-volumes/flume/flume.tar'
+ force: true
+ backup: yes
+
+- name: tar -zxvf flume.tar
+ shell: cd {{ volume_path }}/tsg3.0-volumes/flume/ && tar -zxvf flume.tar && rm -rf flume.tar
+
+- name: load new {{ image_name }} image from {{ image_tar_name }}
+ docker_image:
+ name: '{{ image_name }}'
+ tag: '{{ image_tag_name }}'
+ load_path: '{{ soft_home_path }}/{{ image_tar_name }}'
+ source: load
+ force_tag: yes
+ force_source: yes
+ when: offline_install #离线模式从本地load文件加载
+
+- name: load new {{ image_name }} image from docker-registry
+ docker_image:
+ name: '{{ docker_registry_image_and_tag }}'
+ source: pull
+ force_tag: yes
+ force_source: yes
+ when: offline_install==false #在线模式从docker-registry中pull镜像
+
+- name: copy {{ image_name }} docker-compose.yml
+ template:
+ src: docker-compose.yml.j2
+ dest: '{{ soft_home_path }}/docker-compose.yml'
+ backup: yes
+
+- name: start {{ image_container_name }} container
+ docker_compose:
+ project_src: '{{ soft_home_path }}'
diff --git a/internal-flume/ansible_role/flume/tasks/main.yml b/internal-flume/ansible_role/flume/tasks/main.yml
new file mode 100644
index 0000000..886a51f
--- /dev/null
+++ b/internal-flume/ansible_role/flume/tasks/main.yml
@@ -0,0 +1 @@
+- include: load_flume.yml
diff --git a/internal-flume/ansible_role/flume/templates/docker-compose.yml.j2 b/internal-flume/ansible_role/flume/templates/docker-compose.yml.j2
new file mode 100644
index 0000000..b21a124
--- /dev/null
+++ b/internal-flume/ansible_role/flume/templates/docker-compose.yml.j2
@@ -0,0 +1,40 @@
+version: '2'
+
+services:
+ flume:
+{% if offline_install %}
+ image: {{ image_name }}:{{ image_tag_name }}
+{% else %}
+ image: {{ docker_registry_image_and_tag }}
+{% endif %}
+ restart: always
+ container_name: {{ image_container_name }}
+ environment:
+ MACHINE1: {{ MACHINE1 }}
+ MACHINE2: {{ MACHINE2 }}
+ MACHINE3: {{ MACHINE3 }}
+ MACHINE4: {{ MACHINE4 }}
+ MACHINE5: {{ MACHINE5 }}
+ MACHINE6: {{ MACHINE6 }}
+ MACHINE7: {{ MACHINE7 }}
+ MACHINE8: {{ MACHINE8 }}
+ MACHINE9: {{ MACHINE9 }}
+ MACHINE10: {{ MACHINE10 }}
+ MACHINE11: {{ MACHINE11 }}
+ MACHINE12: {{ MACHINE12 }}
+ MACHINE13: {{ MACHINE13 }}
+ SESPROCESSNUM: 1
+ SESPORTBEGIN: 35000
+ KAFKA_ADDR: {{ kafka_addr }}
+ volumes:
+ - "{{ volume_path }}/tsg3.0-volumes/flume/conf:/home/apache-flume-1.9.0-bin/conf"
+ - "{{ volume_path }}/tsg3.0-volumes/flume/flumeController:/home/apache-flume-1.9.0-bin/flumeController"
+ - "{{ volume_path }}/tsg3.0-volumes/flume/flumeMonitor:/home/apache-flume-1.9.0-bin/flumeMonitor"
+ - "{{ volume_path }}/tsg3.0-volumes/flume/plugins.d:/home/apache-flume-1.9.0-bin/plugins.d"
+ - "{{ volume_path }}/tsg3.0-volumes/flume/dat/:/home/dat/"
+ - "{{ volume_path }}/tsg3.0-volumes/flume/json:/home/json"
+ - "{{ volume_path }}/tsg3.0-volumes/flume/data:/home/data"
+ - "/etc/localtime:/etc/localtime:ro"
+ - "/etc/timezone:/etc/timezone:ro"
+ network_mode: "host"
+
diff --git a/internal-flume/ansible_role/flume/vars/main.yml b/internal-flume/ansible_role/flume/vars/main.yml
new file mode 100644
index 0000000..4408c36
--- /dev/null
+++ b/internal-flume/ansible_role/flume/vars/main.yml
@@ -0,0 +1,39 @@
+#注意这里面定义的key中不能包含减号-,否则程序无法识别定义的key
+
+#########################从docker仓库pull镜像的相关属性##########################################################
+#docker仓库(harbor)中的地址,镜像名称和tag
+docker_registry_image_and_tag: 192.168.40.153:9080/tsg/flume:1.9.0 #bifang_api镜像和tag名称
+
+#########################离线加载docker镜像的相关属性##########################################################
+#离线模式下使用:bifang-api的docker镜像文件名称,在../file中的文件名称
+image_tar_name: flume-1.9.0.tar
+
+
+
+##########################公共属性#############################
+#设置容器的hostname
+#container_hostname: bifang-api
+#{{ image_tar_name }}中的镜像名称,或者{{ docker_registry_image_and_tag }}中的镜像名称,不含版本号
+image_name: flume
+#{{ image_tar_name }}的镜像版本号,或者{{ docker_registry_image_and_tag }}中的镜像版本号
+image_tag_name: 1.9.0
+#{{ image_tar_name }}的镜像,或者{{ docker_registry_image_and_tag }}中的镜像启动后的容器名称
+image_container_name: flume
+#镜像文件,docker-compose文件以及数据卷共享目录所在的位置,注意路径最后面不带/
+soft_home_path: /home/test/flume
+
+MACHINE1: 192.168.60.101
+MACHINE2: 192.168.60.102
+MACHINE3: 192.168.60.103
+MACHINE4: 192.168.60.104
+MACHINE5: 192.168.60.105
+MACHINE6: 192.168.60.106
+MACHINE7: 192.168.60.107
+MACHINE8: 192.168.60.108
+MACHINE9: 192.168.60.109
+MACHINE10: 192.168.60.110
+MACHINE11: 192.168.60.111
+MACHINE12: 192.168.60.112
+MACHINE13: 192.168.60.113
+
+kafka_addr: 192.168.1.3:9093
diff --git a/internal-flume/conf/k2f/k2f.properties b/internal-flume/conf/k2f/k2f.properties
deleted file mode 100644
index 6c08f3a..0000000
--- a/internal-flume/conf/k2f/k2f.properties
+++ /dev/null
@@ -1,22 +0,0 @@
-#为source channel sink起名
-a1.channels = c1
-a1.sinks = k1
-#指定channel为kafka channel,省略source使得效率更高
-a1.channels.c1.type = org.apache.flume.channel.kafka.KafkaChannel
-a1.channels.c1.kafka.bootstrap.servers = 192.168.40.119:9092
-a1.channels.c1.kafka.topic = test
-a1.channels.c1.kafka.consumer.group.id = lxk_0512
-a1.channels.c1.kafka.consumer.auto.offset.reset = latest
-a1.channels.c1.kafka.consumer.enable.auto.commit = true
-a1.channels.c1.kafka.consumer.fetch.max.wait.ms = 1000
-a1.channels.c1.kafka.consumer.fetch.min.bytes = 10485760
-a1.channels.c1.parseAsFlumeEvent = false
-
-#原始file roll sink,将数据写到本机磁盘
-a1.sinks.k1.type = file_roll
-a1.sinks.k1.channel = c1
-a1.sinks.k1.sink.pathManager = default
-a1.sinks.k1.sink.pathManager.extension = dat
-a1.sinks.k1.sink.pathManager.prefix = test-
-a1.sinks.k1.sink.rollInterval = 30
-a1.sinks.k1.sink.directory = /home/test/log
diff --git a/internal-flume/conf/k2f/pro/2f2k.properties b/internal-flume/conf/k2f/pro/2f2k.properties
deleted file mode 100644
index 5e5c573..0000000
--- a/internal-flume/conf/k2f/pro/2f2k.properties
+++ /dev/null
@@ -1,24 +0,0 @@
-a2.sources = s2
-a2.channels = c2
-#指定source收集到的数据发送到哪个管道
-a2.sources.s2.channels = c2
-#taildir source
-a2.sources.s2.type = TAILDIR
-a2.sources.s2.channels = c2
-a2.sources.s2.positionFile = /home/test/2taildir_position.json
-a2.sources.s2.filegroups = f2
-a2.sources.s2.filegroups.f2 = /home/test/log2/.*dat
-a2.sources.s2.maxBatchCount = 1000
-
-# kafka channel充当生产者
-a2.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-a2.channels.c2.brokerList = 192.168.40.222:9092
-a2.channels.c2.zookeeperConnect=192.168.40.222:2181/kafka
-a2.channels.c2.topic = recv_test
-#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
-a2.channels.c2.parseAsFlumeEvent = false
-a2.channels.c2.kafka.producer.acks = 1
-a2.channels.c2.producer.type=async
-a2.channels.c2.queue.buffering.max.ms = 5000
-a2.channels.c2.queue.buffering.max.messages=20000
-a2.channels.c2.batch.num.messages=500
diff --git a/internal-flume/conf/k2f/pro/4f2k.properties b/internal-flume/conf/k2f/pro/4f2k.properties
deleted file mode 100644
index bf0408b..0000000
--- a/internal-flume/conf/k2f/pro/4f2k.properties
+++ /dev/null
@@ -1,58 +0,0 @@
-a2.sources = s2
-a2.channels = c2 c3 c4
-#指定source收集到的数据发送到哪个管道
-a2.sources.s2.channels = c2
-#taildir source
-a2.sources.s2.type = TAILDIR
-a2.sources.s2.positionFile = /home/test/4taildir_position.json
-a2.sources.s2.filegroups = f3
-a2.sources.s2.filegroups.f3 = /home/test/recv/.*dat
-#a2.sources.s2.headers.f3.headerKey1 = channel2
-#a2.sources.s2.filegroups.f4 = /home/test/log/.*dat
-#a2.sources.s2.headers.f4.headerKey1 = channel3
-a2.sources.s2.maxBatchCount = 500
-#a2.sources.s2.selector.type = multiplexing
-
-#channel selector
-#a2.sources.s2.selector.header = headerKey1
-#a2.sources.s2.selector.mapping.channel2=c2
-#a2.sources.s2.selector.mapping.channel3=c3
-
-# kafka channel充当生产者
-a2.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-a2.channels.c2.brokerList = 192.168.40.222:9093
-a2.channels.c2.topic = recv_test
-#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
-a2.channels.c2.parseAsFlumeEvent = false
-a2.channels.c2.kafka.producer.acks = 1
-a2.channels.c2.producer.type=async
-a2.channels.c2.queue.buffering.max.ms = 5000
-a2.channels.c2.queue.buffering.max.messages=200000
-a2.channels.c2.batch.num.messages=5000
-
-#第二个channel
-a2.channels.c3.type = org.apache.flume.channel.kafka.KafkaChannel
-a2.channels.c3.brokerList = 192.168.40.222:9093
-a2.channels.c3.topic = recv_test
-#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
-a2.channels.c3.parseAsFlumeEvent = false
-a2.channels.c3.kafka.producer.acks = 1
-a2.channels.c3.producer.type=async
-a2.channels.c3.queue.buffering.max.ms = 5000
-a2.channels.c3.queue.buffering.max.messages=20000
-a2.channels.c3.batch.num.messages=500
-
-
-
-#第三个channel
-a2.channels.c4.type = org.apache.flume.channel.kafka.KafkaChannel
-a2.channels.c4.brokerList = 192.168.40.222:9093
-a2.channels.c4.topic = recv_test
-#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
-a2.channels.c4.parseAsFlumeEvent = false
-a2.channels.c4.kafka.producer.acks = 1
-a2.channels.c4.producer.type=async
-a2.channels.c4.queue.buffering.max.ms = 5000
-a2.channels.c4.queue.buffering.max.messages=20000
-a2.channels.c4.batch.num.messages=500
-
diff --git a/internal-flume/conf/k2f/pro/f2k.properties b/internal-flume/conf/k2f/pro/f2k.properties
deleted file mode 100644
index 7047e42..0000000
--- a/internal-flume/conf/k2f/pro/f2k.properties
+++ /dev/null
@@ -1,24 +0,0 @@
-a2.sources = s2
-a2.channels = c2
-#指定source收集到的数据发送到哪个管道
-a2.sources.s2.channels = c2
-#taildir source
-a2.sources.s2.type = TAILDIR
-a2.sources.s2.channels = c2
-a2.sources.s2.positionFile = /home/test/taildir_position.json
-a2.sources.s2.filegroups = f0
-a2.sources.s2.filegroups.f0 = /home/test/log/.*dat
-a2.sources.s2.maxBatchCount = 1000
-
-# kafka channel充当生产者
-a2.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-a2.channels.c2.brokerList = 192.168.40.222:9092
-a2.channels.c2.zookeeperConnect=192.168.40.222:2181/kafka
-a2.channels.c2.topic = recv_test
-#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
-a2.channels.c2.parseAsFlumeEvent = false
-a2.channels.c2.kafka.producer.acks = 1
-a2.channels.c2.producer.type=async
-a2.channels.c2.queue.buffering.max.ms = 5000
-a2.channels.c2.queue.buffering.max.messages=20000
-a2.channels.c2.batch.num.messages=500
diff --git a/internal-flume/conf/k2f/pro/f2k1.properties b/internal-flume/conf/k2f/pro/f2k1.properties
deleted file mode 100644
index 2dadced..0000000
--- a/internal-flume/conf/k2f/pro/f2k1.properties
+++ /dev/null
@@ -1,23 +0,0 @@
-a2.sources = s2
-a2.channels = c2
-#指定source收集到的数据发送到哪个管道
-a2.sources.s2.channels = c2
-#taildir source
-a2.sources.s2.type = TAILDIR
-a2.sources.s2.channels = c2
-a2.sources.s2.positionFile = /home/test/1taildir_position.json
-a2.sources.s2.filegroups = f1
-a2.sources.s2.filegroups.f1 = /home/test/log/.*[0-2].dat
-a2.sources.s2.maxBatchCount = 1000
-
-# kafka channel充当生产者
-a2.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-a2.channels.c2.brokerList = 192.168.40.222:9093
-a2.channels.c2.topic = recv_test
-#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
-a2.channels.c2.parseAsFlumeEvent = false
-a2.channels.c2.kafka.producer.acks = 1
-a2.channels.c2.producer.type=async
-a2.channels.c2.queue.buffering.max.ms = 5000
-a2.channels.c2.queue.buffering.max.messages=20000
-a2.channels.c2.batch.num.messages=500
diff --git a/internal-flume/conf/k2f/pro/f2k2.properties b/internal-flume/conf/k2f/pro/f2k2.properties
deleted file mode 100644
index e17c3e2..0000000
--- a/internal-flume/conf/k2f/pro/f2k2.properties
+++ /dev/null
@@ -1,23 +0,0 @@
-a2.sources = s2
-a2.channels = c2
-#指定source收集到的数据发送到哪个管道
-a2.sources.s2.channels = c2
-#taildir source
-a2.sources.s2.type = TAILDIR
-a2.sources.s2.channels = c2
-a2.sources.s2.positionFile = /home/test/2taildir_position.json
-a2.sources.s2.filegroups = f1
-a2.sources.s2.filegroups.f1 = /home/test/log/.*[3-4].dat
-a2.sources.s2.maxBatchCount = 1000
-
-# kafka channel充当生产者
-a2.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-a2.channels.c2.brokerList = 192.168.40.222:9093
-a2.channels.c2.topic = recv_test
-#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
-a2.channels.c2.parseAsFlumeEvent = false
-a2.channels.c2.kafka.producer.acks = 1
-a2.channels.c2.producer.type=async
-a2.channels.c2.queue.buffering.max.ms = 5000
-a2.channels.c2.queue.buffering.max.messages=20000
-a2.channels.c2.batch.num.messages=500
diff --git a/internal-flume/conf/k2f/pro/f2k3.properties b/internal-flume/conf/k2f/pro/f2k3.properties
deleted file mode 100644
index 2dadced..0000000
--- a/internal-flume/conf/k2f/pro/f2k3.properties
+++ /dev/null
@@ -1,23 +0,0 @@
-a2.sources = s2
-a2.channels = c2
-#指定source收集到的数据发送到哪个管道
-a2.sources.s2.channels = c2
-#taildir source
-a2.sources.s2.type = TAILDIR
-a2.sources.s2.channels = c2
-a2.sources.s2.positionFile = /home/test/1taildir_position.json
-a2.sources.s2.filegroups = f1
-a2.sources.s2.filegroups.f1 = /home/test/log/.*[0-2].dat
-a2.sources.s2.maxBatchCount = 1000
-
-# kafka channel充当生产者
-a2.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-a2.channels.c2.brokerList = 192.168.40.222:9093
-a2.channels.c2.topic = recv_test
-#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
-a2.channels.c2.parseAsFlumeEvent = false
-a2.channels.c2.kafka.producer.acks = 1
-a2.channels.c2.producer.type=async
-a2.channels.c2.queue.buffering.max.ms = 5000
-a2.channels.c2.queue.buffering.max.messages=20000
-a2.channels.c2.batch.num.messages=500
diff --git a/internal-flume/conf/k2f/pro/file.properties b/internal-flume/conf/k2f/pro/file.properties
deleted file mode 100644
index f6813e7..0000000
--- a/internal-flume/conf/k2f/pro/file.properties
+++ /dev/null
@@ -1,29 +0,0 @@
-#为source channel sink起名
-a1.sources = s1
-a1.channels = c1
-a1.sinks = k1
-#指定source收集到的数据发送到哪个管道
-a1.sources.s1.channels = c1
-#指定source数据收集策略
-a1.sources.s1.type = org.apache.flume.source.kafka.KafkaSource
-a1.sources.s1.batchSize = 3000
-a1.sources.s1.batchDurationMillis = 100
-a1.sources.s1.kafka.bootstrap.servers = 192.168.40.203:9092
-a1.sources.s1.kafka.topics = test
-a1.sources.s1.kafka.consumer.group.id = lxk_0429
-
-#指定channel为memory,即表示所有的数据都装进memory当中
-a1.channels.c1.type = memory
-a1.channels.c1.capacity = 2000000
-a1.channels.c1.transactionCapacity = 30000
-a1.channels.c1.byteCapacityBufferPercentage = 40
-a1.channels.c1.byteCapacity = 2147483648
-
-#原始file roll sink,将数据写到本机磁盘
-a1.sinks.k1.type = file_roll
-a1.sinks.k1.channel = c1
-a1.sinks.k1.sink.pathManager = default
-a1.sinks.k1.sink.pathManager.extension = dat
-a1.sinks.k1.sink.pathManager.prefix = test-
-a1.sinks.k1.sink.rollInterval = 30
-a1.sinks.k1.sink.directory = /home/test/log2
diff --git a/internal-flume/conf/k2f/pro/file1.properties b/internal-flume/conf/k2f/pro/file1.properties
deleted file mode 100644
index e852981..0000000
--- a/internal-flume/conf/k2f/pro/file1.properties
+++ /dev/null
@@ -1,29 +0,0 @@
-#为source channel sink起名
-a1.sources = s1
-a1.channels = c1
-a1.sinks = k1
-#指定source收集到的数据发送到哪个管道
-a1.sources.s1.channels = c1
-#指定source数据收集策略
-a1.sources.s1.type = org.apache.flume.source.kafka.KafkaSource
-a1.sources.s1.batchSize = 3000
-a1.sources.s1.batchDurationMillis = 100
-a1.sources.s1.kafka.bootstrap.servers = 192.168.40.203:9092
-a1.sources.s1.kafka.topics = test
-a1.sources.s1.kafka.consumer.group.id = lxk_0429
-
-#指定channel为memory,即表示所有的数据都装进memory当中
-a1.channels.c1.type = memory
-a1.channels.c1.capacity = 2000000
-a1.channels.c1.transactionCapacity = 30000
-a1.channels.c1.byteCapacityBufferPercentage = 40
-a1.channels.c1.byteCapacity = 2147483648
-
-#原始file roll sink,将数据写到本机磁盘
-a1.sinks.k1.type = file_roll
-a1.sinks.k1.channel = c1
-a1.sinks.k1.sink.pathManager = default
-a1.sinks.k1.sink.pathManager.extension = dat
-a1.sinks.k1.sink.pathManager.prefix = test-
-a1.sinks.k1.sink.rollInterval = 30
-a1.sinks.k1.sink.directory = /home/test/log1
diff --git a/internal-flume/conf/k2f/pro/file2.properties b/internal-flume/conf/k2f/pro/file2.properties
deleted file mode 100644
index f6813e7..0000000
--- a/internal-flume/conf/k2f/pro/file2.properties
+++ /dev/null
@@ -1,29 +0,0 @@
-#为source channel sink起名
-a1.sources = s1
-a1.channels = c1
-a1.sinks = k1
-#指定source收集到的数据发送到哪个管道
-a1.sources.s1.channels = c1
-#指定source数据收集策略
-a1.sources.s1.type = org.apache.flume.source.kafka.KafkaSource
-a1.sources.s1.batchSize = 3000
-a1.sources.s1.batchDurationMillis = 100
-a1.sources.s1.kafka.bootstrap.servers = 192.168.40.203:9092
-a1.sources.s1.kafka.topics = test
-a1.sources.s1.kafka.consumer.group.id = lxk_0429
-
-#指定channel为memory,即表示所有的数据都装进memory当中
-a1.channels.c1.type = memory
-a1.channels.c1.capacity = 2000000
-a1.channels.c1.transactionCapacity = 30000
-a1.channels.c1.byteCapacityBufferPercentage = 40
-a1.channels.c1.byteCapacity = 2147483648
-
-#原始file roll sink,将数据写到本机磁盘
-a1.sinks.k1.type = file_roll
-a1.sinks.k1.channel = c1
-a1.sinks.k1.sink.pathManager = default
-a1.sinks.k1.sink.pathManager.extension = dat
-a1.sinks.k1.sink.pathManager.prefix = test-
-a1.sinks.k1.sink.rollInterval = 30
-a1.sinks.k1.sink.directory = /home/test/log2
diff --git a/internal-flume/conf/k2f/pro/k2f.properties b/internal-flume/conf/k2f/pro/k2f.properties
deleted file mode 100644
index 35d7d68..0000000
--- a/internal-flume/conf/k2f/pro/k2f.properties
+++ /dev/null
@@ -1,22 +0,0 @@
-#为source channel sink起名
-a1.channels = c1
-a1.sinks = k1
-#指定channel为kafka channel,省略source使得效率更高
-a1.channels.c1.type = org.apache.flume.channel.kafka.KafkaChannel
-a1.channels.c1.kafka.bootstrap.servers = 192.168.40.119:9092
-a1.channels.c1.kafka.topic = test
-a1.channels.c1.kafka.consumer.group.id = lxk_0509
-a1.channels.c1.kafka.consumer.auto.offset.reset = latest
-a1.channels.c1.kafka.consumer.enable.auto.commit = true
-a1.channels.c1.kafka.consumer.fetch.max.wait.ms = 1000
-a1.channels.c1.kafka.consumer.fetch.min.bytes = 10485760
-a1.channels.c1.parseAsFlumeEvent = false
-
-#原始file roll sink,将数据写到本机磁盘
-a1.sinks.k1.type = file_roll
-a1.sinks.k1.channel = c1
-a1.sinks.k1.sink.pathManager = default
-a1.sinks.k1.sink.pathManager.extension = dat
-a1.sinks.k1.sink.pathManager.prefix = test-
-a1.sinks.k1.sink.rollInterval = 60
-a1.sinks.k1.sink.directory = /home/test/log
diff --git a/internal-flume/conf/k2f/pro/kafka.properties b/internal-flume/conf/k2f/pro/kafka.properties
deleted file mode 100644
index da560b2..0000000
--- a/internal-flume/conf/k2f/pro/kafka.properties
+++ /dev/null
@@ -1,25 +0,0 @@
-a2.sources = s2
-a2.channels = c2
-#指定source收集到的数据发送到哪个管道
-a2.sources.s2.channels = c2
-#taildir source
-a2.sources.s2.type = TAILDIR
-a2.sources.s2.channels = c2
-a2.sources.s2.positionFile = /home/test/taildir_position.json
-a2.sources.s2.filegroups = f1
-a2.sources.s2.filegroups.f1 = /home/test/log/.*dat
-a2.sources.s2.filegroups.f2 = /home/test/log2/.*dat
-a2.sources.s2.maxBatchCount = 1000
-
-# kafka channel充当生产者
-a2.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-a2.channels.c2.brokerList = 192.168.40.203:9092
-a2.channels.c2.zookeeperConnect=192.168.40.203:2181/kafka
-a2.channels.c2.topic = recv_test
-#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
-a2.channels.c2.parseAsFlumeEvent = false
-a2.channels.c2.kafka.producer.acks = 1
-a2.channels.c2.producer.type=async
-a2.channels.c2.queue.buffering.max.ms = 5000
-a2.channels.c2.queue.buffering.max.messages=20000
-a2.channels.c2.batch.num.messages=500
diff --git a/internal-flume/conf/k2f/pro/kafka1.properties b/internal-flume/conf/k2f/pro/kafka1.properties
deleted file mode 100644
index ed30b16..0000000
--- a/internal-flume/conf/k2f/pro/kafka1.properties
+++ /dev/null
@@ -1,25 +0,0 @@
-a2.sources = s2
-a2.channels = c2
-#指定source收集到的数据发送到哪个管道
-a2.sources.s2.channels = c2
-#taildir source
-a2.sources.s2.type = TAILDIR
-a2.sources.s2.channels = c2
-a2.sources.s2.positionFile = /var/taildir_position.json
-a2.sources.s2.filegroups = f1
-a2.sources.s2.filegroups.f1 = /home/test/log1/.*dat
-a2.sources.s2.filegroups.f2 = /home/test/log2/.*dat
-a2.sources.s2.maxBatchCount = 1000
-
-# kafka channel充当生产者
-a2.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-a2.channels.c2.brokerList = 192.168.40.203:9092
-a2.channels.c2.zookeeperConnect=192.168.40.203:2181/kafka
-a2.channels.c2.topic = recv_test
-#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
-a2.channels.c2.parseAsFlumeEvent = false
-a2.channels.c2.kafka.producer.acks = 1
-a2.channels.c2.producer.type=async
-a2.channels.c2.queue.buffering.max.ms = 5000
-a2.channels.c2.queue.buffering.max.messages=20000
-a2.channels.c2.batch.num.messages=500
diff --git a/internal-flume/conf/k2f/pro/kafka2.properties b/internal-flume/conf/k2f/pro/kafka2.properties
deleted file mode 100644
index 6040a05..0000000
--- a/internal-flume/conf/k2f/pro/kafka2.properties
+++ /dev/null
@@ -1,25 +0,0 @@
-a3.sources = s2
-a3.channels = c2
-#指定source收集到的数据发送到哪个管道
-a3.sources.s2.channels = c2
-#taildir source
-a3.sources.s2.type = TAILDIR
-a3.sources.s2.channels = c2
-a3.sources.s2.positionFile = /var/taildir2_position.json
-a3.sources.s2.filegroups = f2
-a3.sources.s2.filegroups.f1 = /home/test/log1/.*dat
-a3.sources.s2.filegroups.f2 = /home/test/log2/.*dat
-a3.sources.s2.maxBatchCount = 1000
-
-# kafka channel充当生产者
-a3.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-a3.channels.c2.brokerList = 192.168.40.203:9092
-a3.channels.c2.zookeeperConnect=192.168.40.203:2181/kafka
-a3.channels.c2.topic = recv_test
-#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
-a3.channels.c2.parseAsFlumeEvent = false
-a3.channels.c2.kafka.producer.acks = 1
-a3.channels.c2.producer.type=async
-a3.channels.c2.queue.buffering.max.ms = 5000
-a3.channels.c2.queue.buffering.max.messages=20000
-a3.channels.c2.batch.num.messages=500
diff --git a/internal-flume/conf/k2f/pro/log4j.properties b/internal-flume/conf/k2f/pro/log4j.properties
deleted file mode 100644
index c948bff..0000000
--- a/internal-flume/conf/k2f/pro/log4j.properties
+++ /dev/null
@@ -1,68 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-#
-
-# Define some default values that can be overridden by system properties.
-#
-# For testing, it may also be convenient to specify
-# -Dflume.root.logger=DEBUG,console when launching flume.
-
-#flume.root.logger=DEBUG,console
-flume.root.logger=INFO,LOGFILE
-flume.log.dir=./logs
-flume.log.file=flume.log
-
-log4j.logger.org.apache.flume.lifecycle = INFO
-log4j.logger.org.jboss = WARN
-log4j.logger.org.mortbay = INFO
-log4j.logger.org.apache.avro.ipc.NettyTransceiver = WARN
-log4j.logger.org.apache.hadoop = INFO
-log4j.logger.org.apache.hadoop.hive = ERROR
-
-# Define the root logger to the system property "flume.root.logger".
-log4j.rootLogger=${flume.root.logger}
-
-
-# Stock log4j rolling file appender
-# Default log rotation configuration
-log4j.appender.LOGFILE=org.apache.log4j.RollingFileAppender
-log4j.appender.LOGFILE.MaxFileSize=100MB
-log4j.appender.LOGFILE.MaxBackupIndex=10
-log4j.appender.LOGFILE.File=${flume.log.dir}/${flume.log.file}
-log4j.appender.LOGFILE.layout=org.apache.log4j.PatternLayout
-log4j.appender.LOGFILE.layout.ConversionPattern=%d{dd MMM yyyy HH:mm:ss,SSS} %-5p [%t] (%C.%M:%L) %x - %m%n
-
-
-# Warning: If you enable the following appender it will fill up your disk if you don't have a cleanup job!
-# This uses the updated rolling file appender from log4j-extras that supports a reliable time-based rolling policy.
-# See http://logging.apache.org/log4j/companions/extras/apidocs/org/apache/log4j/rolling/TimeBasedRollingPolicy.html
-# Add "DAILY" to flume.root.logger above if you want to use this
-log4j.appender.DAILY=org.apache.log4j.rolling.RollingFileAppender
-log4j.appender.DAILY.rollingPolicy=org.apache.log4j.rolling.TimeBasedRollingPolicy
-log4j.appender.DAILY.rollingPolicy.ActiveFileName=${flume.log.dir}/${flume.log.file}
-log4j.appender.DAILY.rollingPolicy.FileNamePattern=${flume.log.dir}/${flume.log.file}.%d{yyyy-MM-dd}
-log4j.appender.DAILY.layout=org.apache.log4j.PatternLayout
-log4j.appender.DAILY.layout.ConversionPattern=%d{dd MMM yyyy HH:mm:ss,SSS} %-5p [%t] (%C.%M:%L) %x - %m%n
-
-
-# console
-# Add "console" to flume.root.logger above if you want to use this
-log4j.appender.console=org.apache.log4j.ConsoleAppender
-log4j.appender.console.target=System.err
-log4j.appender.console.layout=org.apache.log4j.PatternLayout
-log4j.appender.console.layout.ConversionPattern=%d (%t) [%p - %l] %m%n
diff --git a/internal-flume/conf/k2f/pro/tail.properties b/internal-flume/conf/k2f/pro/tail.properties
deleted file mode 100644
index c677318..0000000
--- a/internal-flume/conf/k2f/pro/tail.properties
+++ /dev/null
@@ -1,34 +0,0 @@
-a2.sources = s2
-a2.channels = c2
-a2.sinks = k2
-#指定source收集到的数据发送到哪个管道
-a2.sources.s2.channels = c2
-
-a2.sources.s2.type = TAILDIR
-a2.sources.s2.channels = c2
-a2.sources.s2.positionFile = /home/test/taildir_position.json
-a2.sources.s2.filegroups = f1 f2
-a2.sources.s2.filegroups.f1 = /home/test/log1/.*dat
-a2.sources.s2.headers.f1.headerKey1 = value1
-a2.sources.s2.filegroups.f2 = /home/test/log2/.*dat
-a2.sources.s2.headers.f2.headerKey1 = value2
-a2.sources.s2.headers.f2.headerKey2 = value2-2
-a2.sources.s2.fileHeader = true
-a2.sources.s2.maxBatchCount = 1000
-
-
-#指定channel为memory,即表示所有的数据都装进memory当中
-a2.channels.c2.type = memory
-a2.channels.c2.capacity = 2000000
-a2.channels.c2.transactionCapacity = 30000
-a2.channels.c2.byteCapacityBufferPercentage = 40
-a2.channels.c2.byteCapacity = 2147483648
-
-#原始file roll sink,将数据写到本机磁盘
-a2.sinks.k2.type = file_roll
-a2.sinks.k2.channel = c2
-a2.sinks.k2.sink.pathManager = default
-a2.sinks.k2.sink.pathManager.extension = dat
-a2.sinks.k2.sink.pathManager.prefix = recv-
-a2.sinks.k2.sink.rollInterval = 60
-a2.sinks.k2.sink.directory = /home/test/recv1
diff --git a/internal-flume/conf/k2f/restart_log/restart_f2k.log b/internal-flume/conf/k2f/restart_log/restart_f2k.log
deleted file mode 100644
index 9d5fe99..0000000
--- a/internal-flume/conf/k2f/restart_log/restart_f2k.log
+++ /dev/null
@@ -1,213 +0,0 @@
-##############################################################
-20200509174727 ---> the f2k_APP restart ---> 0
-20200509174737 ---> the f2k_APP restart ---> 1
-##############################################################
-##############################################################
-##############################################################
-##############################################################
-##############################################################
-20200511183420 ---> the f2k_APP restart ---> 1
-20200511183430 ---> the f2k_APP restart ---> 2
-##############################################################
-20200511183639 ---> the f2k_APP restart ---> 1
-20200511183649 ---> the f2k_APP restart ---> 2
-20200511183659 ---> the f2k_APP restart ---> 3
-20200511183709 ---> the f2k_APP restart ---> 4
-20200511183719 ---> the f2k_APP restart ---> 5
-20200511183729 ---> the f2k_APP restart ---> 6
-20200511183739 ---> the f2k_APP restart ---> 7
-20200511183749 ---> the f2k_APP restart ---> 8
-20200511183759 ---> the f2k_APP restart ---> 9
-20200511183809 ---> the f2k_APP restart ---> 10
-20200511183819 ---> the f2k_APP restart ---> 11
-20200511183829 ---> the f2k_APP restart ---> 12
-20200511183839 ---> the f2k_APP restart ---> 13
-20200511183849 ---> the f2k_APP restart ---> 14
-20200511183859 ---> the f2k_APP restart ---> 15
-20200511183909 ---> the f2k_APP restart ---> 16
-20200511183919 ---> the f2k_APP restart ---> 17
-20200511183930 ---> the f2k_APP restart ---> 18
-20200511183940 ---> the f2k_APP restart ---> 19
-20200511183950 ---> the f2k_APP restart ---> 20
-20200511184000 ---> the f2k_APP restart ---> 21
-20200511184010 ---> the f2k_APP restart ---> 22
-20200511184020 ---> the f2k_APP restart ---> 23
-20200511184030 ---> the f2k_APP restart ---> 24
-20200511184040 ---> the f2k_APP restart ---> 25
-20200511184050 ---> the f2k_APP restart ---> 26
-20200511184100 ---> the f2k_APP restart ---> 27
-20200511184110 ---> the f2k_APP restart ---> 28
-20200511184120 ---> the f2k_APP restart ---> 29
-20200511184130 ---> the f2k_APP restart ---> 30
-20200511184140 ---> the f2k_APP restart ---> 31
-20200511184150 ---> the f2k_APP restart ---> 32
-20200511184200 ---> the f2k_APP restart ---> 33
-20200511184210 ---> the f2k_APP restart ---> 34
-20200511184220 ---> the f2k_APP restart ---> 35
-20200511184230 ---> the f2k_APP restart ---> 36
-20200511184240 ---> the f2k_APP restart ---> 37
-20200511184250 ---> the f2k_APP restart ---> 38
-20200511184300 ---> the f2k_APP restart ---> 39
-20200511184310 ---> the f2k_APP restart ---> 40
-20200511184321 ---> the f2k_APP restart ---> 41
-20200511184331 ---> the f2k_APP restart ---> 42
-20200511184341 ---> the f2k_APP restart ---> 43
-20200511184351 ---> the f2k_APP restart ---> 44
-20200511184401 ---> the f2k_APP restart ---> 45
-20200511184411 ---> the f2k_APP restart ---> 46
-20200511184421 ---> the f2k_APP restart ---> 47
-20200511184431 ---> the f2k_APP restart ---> 48
-20200511184441 ---> the f2k_APP restart ---> 49
-20200511184451 ---> the f2k_APP restart ---> 50
-20200511184501 ---> the f2k_APP restart ---> 51
-20200511184511 ---> the f2k_APP restart ---> 52
-20200511184521 ---> the f2k_APP restart ---> 53
-20200511184531 ---> the f2k_APP restart ---> 54
-20200511184541 ---> the f2k_APP restart ---> 55
-20200511184551 ---> the f2k_APP restart ---> 56
-20200511184601 ---> the f2k_APP restart ---> 57
-20200511184611 ---> the f2k_APP restart ---> 58
-20200511184621 ---> the f2k_APP restart ---> 59
-20200511184631 ---> the f2k_APP restart ---> 60
-20200511184641 ---> the f2k_APP restart ---> 61
-20200511184651 ---> the f2k_APP restart ---> 62
-20200511184701 ---> the f2k_APP restart ---> 63
-20200511184711 ---> the f2k_APP restart ---> 64
-20200511184721 ---> the f2k_APP restart ---> 65
-20200511184732 ---> the f2k_APP restart ---> 66
-20200511184742 ---> the f2k_APP restart ---> 67
-20200511184752 ---> the f2k_APP restart ---> 68
-20200511184802 ---> the f2k_APP restart ---> 69
-20200511184812 ---> the f2k_APP restart ---> 70
-20200511184822 ---> the f2k_APP restart ---> 71
-##############################################################
-20200511185311 ---> the f2k_APP restart ---> 1
-##############################################################
-20200511185421 ---> the f2k_APP restart ---> 1
-##############################################################
-20200511185532 ---> the f2k_APP restart ---> 1
-20200511185542 ---> the f2k_APP restart ---> 1
-20200511185552 ---> the f2k_APP restart ---> 1
-##############################################################
-20200511185706 ---> the f2k_APP restart ---> 1
-20200511185716 ---> the f2k_APP restart ---> 1+1
-20200511185726 ---> the f2k_APP restart ---> 2+1
-##############################################################
-20200511185837 ---> the f2k_APP restart ---> 1
-##############################################################
-20200511185938 ---> the f2k_APP restart ---> 1
-##############################################################
-20200511190054 ---> the f2k_APP restart ---> 1
-20200511190104 ---> the f2k_APP restart ---> 1+1
-##############################################################
-20200511190604 ---> the f2k_APP restart ---> 1
-20200511190614 ---> the f2k_APP restart ---> 0
-20200511190624 ---> the f2k_APP restart ---> 1
-20200511190634 ---> the f2k_APP restart ---> 0
-##############################################################
-20200511190729 ---> the f2k_APP restart ---> 1
-20200511190739 ---> the f2k_APP restart ---> 1
-20200511190749 ---> the f2k_APP restart ---> 1
-##############################################################
-20200511190950 ---> the f2k_APP restart ---> 1
-20200511191000 ---> the f2k_APP restart ---> 2
-20200511191010 ---> the f2k_APP restart ---> 1
-##############################################################
-20200511191834 ---> the f2k_APP restart ---> 1
-20200511191844 ---> the f2k_APP restart ---> 2
-##############################################################
-20200511192013 ---> the f2k_APP restart ---> 1
-##############################################################
-20200511192134 ---> the f2k_APP restart ---> 1
-20200511192144 ---> the f2k_APP restart ---> 2
-20200511192154 ---> the f2k_APP restart ---> 1
-20200511192204 ---> the f2k_APP restart ---> 2
-##############################################################
-##############################################################
-##############################################################
-##############################################################
-20200511193630 ---> the f2k_APP restart ---> 1
-20200511193640 ---> the f2k_APP restart ---> 2
-##############################################################
-20200512101249 ---> the f2k_APP restart ---> 1
-20200512101259 ---> the f2k_APP restart ---> 2
-20200512101309 ---> the f2k_APP restart ---> 1
-##############################################################
-20200512101433 ---> the f2k_APP restart ---> 1
-##############################################################
-20200512101537 ---> the f2k_APP restart ---> 1
-20200512101547 ---> the f2k_APP restart ---> 2
-20200512101557 ---> the f2k_APP restart ---> 3
-20200512101707 ---> the f2k_APP restart ---> 1
-##############################################################
-20200512102643 ---> the f2k_APP restart ---> 1
-20200512102653 ---> the f2k_APP restart ---> 2
-##############################################################
-20200512102723 ---> the f2k_APP restart ---> 1
-20200512102733 ---> the f2k_APP restart ---> 2
-20200512102743 ---> the f2k_APP restart ---> 3
-##############################################################
-20200512102936 ---> the f2k_APP restart ---> 1
-20200512102946 ---> the f2k_APP restart ---> 2
-20200512102956 ---> the f2k_APP restart ---> 3
-##############################################################
-20200512103250 ---> the f2k_APP restart ---> 1
-20200512103300 ---> the f2k_APP restart ---> 2
-20200512103310 ---> the f2k_APP restart ---> 3
-##############################################################
-20200512113819 ---> the f2k_APP restart ---> 1
-20200512113829 ---> the f2k_APP restart ---> 2
-20200512113839 ---> the f2k_APP restart ---> 3
-##############################################################
-##############################################################
-20200512114211 ---> the f2k_APP restart ---> 1
-20200512114241 ---> the f2k_APP restart ---> 2
-##############################################################
-20200512114550 ---> the f2k_APP restart ---> 1
-20200512114600 ---> the f2k_APP restart ---> 2
-20200512114610 ---> the f2k_APP restart ---> 3
-##############################################################
-##############################################################
-20200512115341 ---> the f2k_APP restart ---> 1
-20200512115351 ---> the f2k_APP restart ---> 2
-20200512115401 ---> the f2k_APP restart ---> 3
-20200512115452 ---> the f2k_APP restart ---> 1
-20200512115502 ---> the f2k_APP restart ---> 2
-20200512115512 ---> the f2k_APP restart ---> 3
-20200512115712 ---> the f2k_APP restart ---> 1
-20200512115722 ---> the f2k_APP restart ---> 2
-20200512115732 ---> the f2k_APP restart ---> 3
-##############################################################
-20200512115933 ---> the f2k_APP restart ---> 1
-20200512115943 ---> the f2k_APP restart ---> 2
-20200512115953 ---> the f2k_APP restart ---> 3
-##############################################################
-20200512134239 ---> the f2k_APP restart ---> 1
-20200512134249 ---> the f2k_APP restart ---> 2
-20200512134259 ---> the f2k_APP restart ---> 3
-20200512134509 ---> the f2k_APP restart ---> 1
-20200512134519 ---> the f2k_APP restart ---> 2
-20200512134530 ---> the f2k_APP restart ---> 3
-20200512140003 ---> the f2k_APP restart ---> 1
-20200512140013 ---> the f2k_APP restart ---> 2
-20200512141056 ---> the f2k_APP restart ---> 1
-20200512141106 ---> the f2k_APP restart ---> 2
-20200512141116 ---> the f2k_APP restart ---> 3
-20200512142500 ---> the f2k_APP restart ---> 1
-20200512142510 ---> the f2k_APP restart ---> 2
-20200512142941 ---> the f2k_APP restart ---> 1
-20200512142951 ---> the f2k_APP restart ---> 2
-20200512143001 ---> the f2k_APP restart ---> 3
-20200512143031 ---> the f2k_APP restart ---> 1
-20200512143041 ---> the f2k_APP restart ---> 2
-20200512143051 ---> the f2k_APP restart ---> 3
-20200512144224 ---> the f2k_APP restart ---> 1
-20200512144234 ---> the f2k_APP restart ---> 2
-20200512144314 ---> the f2k_APP restart ---> 1
-20200512144324 ---> the f2k_APP restart ---> 2
-20200512144355 ---> the f2k_APP restart ---> 1
-20200512144405 ---> the f2k_APP restart ---> 2
-20200512144415 ---> the f2k_APP restart ---> 3
-20200512144635 ---> the f2k_APP restart ---> 1
-20200512144645 ---> the f2k_APP restart ---> 2
-20200512144655 ---> the f2k_APP restart ---> 3
diff --git a/internal-flume/conf/k2f/restart_log/restart_security_f2k.log b/internal-flume/conf/k2f/restart_log/restart_security_f2k.log
deleted file mode 100644
index 690385d..0000000
--- a/internal-flume/conf/k2f/restart_log/restart_security_f2k.log
+++ /dev/null
@@ -1,134 +0,0 @@
-##############################################################
-20200512180055 ---> the security_f2k_APP restart ---> 1
-20200512180105 ---> the security_f2k_APP restart ---> 2
-##############################################################
-20200512180108 ---> the security_f2k_APP restart ---> 1
-20200512180115 ---> the security_f2k_APP restart ---> 3
-20200512180118 ---> the security_f2k_APP restart ---> 2
-20200512180125 ---> the security_f2k_APP restart ---> 1
-20200512180128 ---> the security_f2k_APP restart ---> 3
-20200512180135 ---> the security_f2k_APP restart ---> 2
-20200512180138 ---> the security_f2k_APP restart ---> 1
-20200512180145 ---> the security_f2k_APP restart ---> 3
-20200512180148 ---> the security_f2k_APP restart ---> 2
-20200512180155 ---> the security_f2k_APP restart ---> 1
-20200512180158 ---> the security_f2k_APP restart ---> 3
-20200512180206 ---> the security_f2k_APP restart ---> 2
-20200512180208 ---> the security_f2k_APP restart ---> 1
-20200512180216 ---> the security_f2k_APP restart ---> 3
-20200512180218 ---> the security_f2k_APP restart ---> 2
-20200512180226 ---> the security_f2k_APP restart ---> 1
-20200512180228 ---> the security_f2k_APP restart ---> 3
-20200512180236 ---> the security_f2k_APP restart ---> 2
-20200512180238 ---> the security_f2k_APP restart ---> 1
-20200512180246 ---> the security_f2k_APP restart ---> 3
-20200512180248 ---> the security_f2k_APP restart ---> 2
-20200512180256 ---> the security_f2k_APP restart ---> 1
-20200512180258 ---> the security_f2k_APP restart ---> 3
-##############################################################
-20200512180513 ---> the security_f2k_APP restart ---> 1
-20200512180523 ---> the security_f2k_APP restart ---> 2
-20200512180533 ---> the security_f2k_APP restart ---> 3
-##############################################################
-20200512180536 ---> the security_f2k_APP restart ---> 1
-20200512180546 ---> the security_f2k_APP restart ---> 2
-20200512180556 ---> the security_f2k_APP restart ---> 3
-20200512180626 ---> the security_f2k_APP restart ---> 1
-20200512180637 ---> the security_f2k_APP restart ---> 2
-20200512180707 ---> the security_f2k_APP restart ---> 1
-20200512180717 ---> the security_f2k_APP restart ---> 2
-20200512180747 ---> the security_f2k_APP restart ---> 1
-20200512180757 ---> the security_f2k_APP restart ---> 2
-20200512180827 ---> the security_f2k_APP restart ---> 1
-20200512180837 ---> the security_f2k_APP restart ---> 2
-20200512180907 ---> the security_f2k_APP restart ---> 1
-20200512180917 ---> the security_f2k_APP restart ---> 2
-20200512180947 ---> the security_f2k_APP restart ---> 1
-20200512180957 ---> the security_f2k_APP restart ---> 2
-20200512181028 ---> the security_f2k_APP restart ---> 1
-20200512181038 ---> the security_f2k_APP restart ---> 2
-20200512181108 ---> the security_f2k_APP restart ---> 1
-20200512181118 ---> the security_f2k_APP restart ---> 2
-20200512181148 ---> the security_f2k_APP restart ---> 1
-20200512181158 ---> the security_f2k_APP restart ---> 2
-20200512181228 ---> the security_f2k_APP restart ---> 1
-20200512181238 ---> the security_f2k_APP restart ---> 2
-20200512181308 ---> the security_f2k_APP restart ---> 1
-20200512181318 ---> the security_f2k_APP restart ---> 2
-20200512181348 ---> the security_f2k_APP restart ---> 1
-20200512181358 ---> the security_f2k_APP restart ---> 2
-20200512181429 ---> the security_f2k_APP restart ---> 1
-20200512181439 ---> the security_f2k_APP restart ---> 2
-20200512181449 ---> the security_f2k_APP restart ---> 3
-20200512181519 ---> the security_f2k_APP restart ---> 1
-20200512181529 ---> the security_f2k_APP restart ---> 2
-20200512181539 ---> the security_f2k_APP restart ---> 3
-20200512181609 ---> the security_f2k_APP restart ---> 1
-20200512181619 ---> the security_f2k_APP restart ---> 2
-20200512181649 ---> the security_f2k_APP restart ---> 1
-20200512181659 ---> the security_f2k_APP restart ---> 2
-20200512181729 ---> the security_f2k_APP restart ---> 1
-20200512181739 ---> the security_f2k_APP restart ---> 2
-20200512181809 ---> the security_f2k_APP restart ---> 1
-20200512181820 ---> the security_f2k_APP restart ---> 2
-20200512181850 ---> the security_f2k_APP restart ---> 1
-20200512181900 ---> the security_f2k_APP restart ---> 2
-20200512181910 ---> the security_f2k_APP restart ---> 3
-20200512181920 ---> the security_f2k_APP restart ---> 1
-20200512181930 ---> the security_f2k_APP restart ---> 2
-20200512181940 ---> the security_f2k_APP restart ---> 3
-20200512181950 ---> the security_f2k_APP restart ---> 1
-20200512182000 ---> the security_f2k_APP restart ---> 2
-20200512182010 ---> the security_f2k_APP restart ---> 3
-20200512182020 ---> the security_f2k_APP restart ---> 1
-20200512182030 ---> the security_f2k_APP restart ---> 2
-20200512182040 ---> the security_f2k_APP restart ---> 3
-20200512182050 ---> the security_f2k_APP restart ---> 1
-20200512182100 ---> the security_f2k_APP restart ---> 2
-20200512182110 ---> the security_f2k_APP restart ---> 3
-20200512182120 ---> the security_f2k_APP restart ---> 1
-20200512182130 ---> the security_f2k_APP restart ---> 2
-20200512182140 ---> the security_f2k_APP restart ---> 3
-20200512182150 ---> the security_f2k_APP restart ---> 1
-20200512182200 ---> the security_f2k_APP restart ---> 2
-20200512182211 ---> the security_f2k_APP restart ---> 3
-##############################################################
-20200512183204 ---> the security_f2k_APP restart ---> 1
-20200512183215 ---> the security_f2k_APP restart ---> 2
-20200512183225 ---> the security_f2k_APP restart ---> 3
-20200512183255 ---> the security_f2k_APP restart ---> 1
-20200512183305 ---> the security_f2k_APP restart ---> 2
-20200512183335 ---> the security_f2k_APP restart ---> 1
-20200512183345 ---> the security_f2k_APP restart ---> 2
-20200512183415 ---> the security_f2k_APP restart ---> 1
-20200512183425 ---> the security_f2k_APP restart ---> 2
-20200512183455 ---> the security_f2k_APP restart ---> 1
-20200512183505 ---> the security_f2k_APP restart ---> 2
-20200512183535 ---> the security_f2k_APP restart ---> 1
-20200512183545 ---> the security_f2k_APP restart ---> 2
-20200512183615 ---> the security_f2k_APP restart ---> 1
-20200512183626 ---> the security_f2k_APP restart ---> 2
-20200512183656 ---> the security_f2k_APP restart ---> 1
-20200512183706 ---> the security_f2k_APP restart ---> 2
-20200512183716 ---> the security_f2k_APP restart ---> 3
-20200512183726 ---> the security_f2k_APP restart ---> 1
-20200512183736 ---> the security_f2k_APP restart ---> 2
-20200512183746 ---> the security_f2k_APP restart ---> 3
-20200512183756 ---> the security_f2k_APP restart ---> 1
-20200512183806 ---> the security_f2k_APP restart ---> 2
-20200512183836 ---> the security_f2k_APP restart ---> 1
-20200512183846 ---> the security_f2k_APP restart ---> 2
-20200512183916 ---> the security_f2k_APP restart ---> 1
-20200512183926 ---> the security_f2k_APP restart ---> 2
-20200512183936 ---> the security_f2k_APP restart ---> 3
-20200512184006 ---> the security_f2k_APP restart ---> 1
-20200512184016 ---> the security_f2k_APP restart ---> 2
-20200512184027 ---> the security_f2k_APP restart ---> 3
-20200512184037 ---> the security_f2k_APP restart ---> 1
-20200512184047 ---> the security_f2k_APP restart ---> 2
-20200512184057 ---> the security_f2k_APP restart ---> 3
-20200512184107 ---> the security_f2k_APP restart ---> 1
-20200512184117 ---> the security_f2k_APP restart ---> 2
-20200512184127 ---> the security_f2k_APP restart ---> 3
-20200512184157 ---> the security_f2k_APP restart ---> 1
-20200512184207 ---> the security_f2k_APP restart ---> 2
diff --git a/internal-flume/conf/k2f/start_flume.sh b/internal-flume/conf/k2f/start_flume.sh
deleted file mode 100755
index 619f9aa..0000000
--- a/internal-flume/conf/k2f/start_flume.sh
+++ /dev/null
@@ -1,5 +0,0 @@
-#!/bin/sh
-
-BASE_DIR=$(cd $(dirname $0); pwd)
-
-nohup ${BASE_DIR}/dae_f2k.sh >/dev/null 2>&1 &
diff --git a/internal-flume/conf/proxy/restart_log/restart_proxy_f2k.log b/internal-flume/conf/proxy/restart_log/restart_proxy_f2k.log
deleted file mode 100644
index 47cb5aa..0000000
--- a/internal-flume/conf/proxy/restart_log/restart_proxy_f2k.log
+++ /dev/null
@@ -1,34 +0,0 @@
-##############################################################
-20200512174751 ---> the proxy_f2k_APP restart ---> 1
-20200512174801 ---> the proxy_f2k_APP restart ---> 2
-20200512174811 ---> the proxy_f2k_APP restart ---> 3
-20200512174821 ---> the proxy_f2k_APP restart ---> 4
-20200512174831 ---> the proxy_f2k_APP restart ---> 5
-##############################################################
-20200512175511 ---> the proxy_f2k_APP restart ---> 1
-20200512175521 ---> the proxy_f2k_APP restart ---> 2
-20200512175531 ---> the proxy_f2k_APP restart ---> 3
-20200512175541 ---> the proxy_f2k_APP restart ---> 4
-20200512175551 ---> the proxy_f2k_APP restart ---> 5
-##############################################################
-20200512180513 ---> the proxy_f2k_APP restart ---> 1
-20200512180523 ---> the proxy_f2k_APP restart ---> 2
-20200512180533 ---> the proxy_f2k_APP restart ---> 3
-##############################################################
-20200512180536 ---> the proxy_f2k_APP restart ---> 1
-20200512180546 ---> the proxy_f2k_APP restart ---> 2
-20200512180556 ---> the proxy_f2k_APP restart ---> 3
-20200512180606 ---> the proxy_f2k_APP restart ---> 4
-20200512180616 ---> the proxy_f2k_APP restart ---> 5
-##############################################################
-20200512183204 ---> the proxy_f2k_APP restart ---> 1
-20200512183214 ---> the proxy_f2k_APP restart ---> 2
-20200512183224 ---> the proxy_f2k_APP restart ---> 3
-20200512183234 ---> the proxy_f2k_APP restart ---> 4
-20200512183245 ---> the proxy_f2k_APP restart ---> 5
-##############################################################
-20200512184503 ---> the proxy_f2k_APP restart ---> 1
-20200512184513 ---> the proxy_f2k_APP restart ---> 2
-20200512184523 ---> the proxy_f2k_APP restart ---> 3
-20200512184533 ---> the proxy_f2k_APP restart ---> 4
-20200512184543 ---> the proxy_f2k_APP restart ---> 5
diff --git a/internal-flume/conf/proxy/start_flume.sh b/internal-flume/conf/proxy/start_flume.sh
deleted file mode 100755
index 619f9aa..0000000
--- a/internal-flume/conf/proxy/start_flume.sh
+++ /dev/null
@@ -1,5 +0,0 @@
-#!/bin/sh
-
-BASE_DIR=$(cd $(dirname $0); pwd)
-
-nohup ${BASE_DIR}/dae_f2k.sh >/dev/null 2>&1 &
diff --git a/internal-flume/conf/radius/restart_log/restart_radius_f2k.log b/internal-flume/conf/radius/restart_log/restart_radius_f2k.log
deleted file mode 100644
index ce1f7c9..0000000
--- a/internal-flume/conf/radius/restart_log/restart_radius_f2k.log
+++ /dev/null
@@ -1,28 +0,0 @@
-##############################################################
-20200512173126 ---> the radius_f2k_APP restart ---> 1
-20200512173136 ---> the radius_f2k_APP restart ---> 2
-20200512173146 ---> the radius_f2k_APP restart ---> 3
-20200512173156 ---> the radius_f2k_APP restart ---> 4
-20200512173206 ---> the radius_f2k_APP restart ---> 5
-##############################################################
-20200512180513 ---> the radius_f2k_APP restart ---> 1
-20200512180523 ---> the radius_f2k_APP restart ---> 2
-20200512180533 ---> the radius_f2k_APP restart ---> 3
-##############################################################
-20200512180536 ---> the radius_f2k_APP restart ---> 1
-20200512180546 ---> the radius_f2k_APP restart ---> 2
-20200512180556 ---> the radius_f2k_APP restart ---> 3
-20200512180606 ---> the radius_f2k_APP restart ---> 4
-20200512180616 ---> the radius_f2k_APP restart ---> 5
-##############################################################
-20200512183204 ---> the radius_f2k_APP restart ---> 1
-20200512183214 ---> the radius_f2k_APP restart ---> 2
-20200512183224 ---> the radius_f2k_APP restart ---> 3
-20200512183235 ---> the radius_f2k_APP restart ---> 4
-20200512183245 ---> the radius_f2k_APP restart ---> 5
-##############################################################
-20200512184503 ---> the radius_f2k_APP restart ---> 1
-20200512184513 ---> the radius_f2k_APP restart ---> 2
-20200512184523 ---> the radius_f2k_APP restart ---> 3
-20200512184533 ---> the radius_f2k_APP restart ---> 4
-20200512184543 ---> the radius_f2k_APP restart ---> 5
diff --git a/internal-flume/conf/radius/start_flume.sh b/internal-flume/conf/radius/start_flume.sh
deleted file mode 100755
index 619f9aa..0000000
--- a/internal-flume/conf/radius/start_flume.sh
+++ /dev/null
@@ -1,5 +0,0 @@
-#!/bin/sh
-
-BASE_DIR=$(cd $(dirname $0); pwd)
-
-nohup ${BASE_DIR}/dae_f2k.sh >/dev/null 2>&1 &
diff --git a/internal-flume/conf/security/start_flume.sh b/internal-flume/conf/security/start_flume.sh
deleted file mode 100755
index 619f9aa..0000000
--- a/internal-flume/conf/security/start_flume.sh
+++ /dev/null
@@ -1,5 +0,0 @@
-#!/bin/sh
-
-BASE_DIR=$(cd $(dirname $0); pwd)
-
-nohup ${BASE_DIR}/dae_f2k.sh >/dev/null 2>&1 &
diff --git a/internal-flume/origin/conf/active/active_defence_f2k1.properties b/internal-flume/origin/conf/active/active_defence_f2k1.properties
new file mode 100644
index 0000000..7d8582c
--- /dev/null
+++ b/internal-flume/origin/conf/active/active_defence_f2k1.properties
@@ -0,0 +1,40 @@
+active_denfence.sources = s2
+active_denfence.channels = c2
+#指定source收集到的数据发送到哪个管道
+active_denfence.sources.s2.channels = c2
+#taildir source
+active_denfence.sources.s2.type = TAILDIR
+active_denfence.sources.s2.channels = c2
+#需要映射到宿主机做持久化
+active_denfence.sources.s2.positionFile = /home/test/1active_denfence_position.json
+#需要分组13台主机的数据
+active_denfence.sources.s2.filegroups = f1 f2 f3 f4 f5 f6 f7 f8 f9 f10 f11 f12 f13
+#需要分组按照序号 [0-9] 目前最多支持10个进程,可以增加
+active_denfence.sources.s2.filegroups.f1 = /home/data/192.168.60.101/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f2 = /home/data/192.168.60.102/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f3 = /home/data/192.168.60.103/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f4 = /home/data/192.168.60.104/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f5 = /home/data/192.168.60.105/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f6 = /home/data/192.168.60.106/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f7 = /home/data/192.168.60.107/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f8 = /home/data/192.168.60.108/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f9 = /home/data/192.168.60.109/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f10 = /home/data/192.168.60.110/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f11 = /home/data/192.168.60.111/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f12 = /home/data/192.168.60.112/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+active_denfence.sources.s2.filegroups.f13 = /home/data/192.168.60.113/ACTIVE-DEFENCE-EVENT-LOG/.*[0-1].dat
+
+
+active_denfence.sources.s2.maxBatchCount = 10000
+
+# kafka channel充当生产者
+active_denfence.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
+active_denfence.channels.c2.brokerList = 192.168.1.3:9093
+active_denfence.channels.c2.topic = ACTIVE-DEFENCE-EVENT-LOG
+#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
+active_denfence.channels.c2.parseAsFlumeEvent = false
+active_denfence.channels.c2.kafka.producer.acks = 1
+active_denfence.channels.c2.producer.type=sync
+active_denfence.channels.c2.queue.buffering.max.ms = 5000
+active_denfence.channels.c2.queue.buffering.max.messages=20000
+active_denfence.channels.c2.batch.num.messages=5000
diff --git a/internal-flume/origin/conf/active/active_defence_f2k2.properties b/internal-flume/origin/conf/active/active_defence_f2k2.properties
new file mode 100644
index 0000000..c0dd97e
--- /dev/null
+++ b/internal-flume/origin/conf/active/active_defence_f2k2.properties
@@ -0,0 +1,40 @@
+active_denfence.sources = s2
+active_denfence.channels = c2
+#指定source收集到的数据发送到哪个管道
+active_denfence.sources.s2.channels = c2
+#taildir source
+active_denfence.sources.s2.type = TAILDIR
+active_denfence.sources.s2.channels = c2
+#需要映射到宿主机做持久化
+active_denfence.sources.s2.positionFile = /home/test/2active_denfence_position.json
+#需要分组13台主机的数据
+active_denfence.sources.s2.filegroups = f1 f2 f3 f4 f5 f6 f7 f8 f9 f10 f11 f12 f13
+#需要分组按照序号 [0-9] 目前最多支持10个进程,可以增加
+active_denfence.sources.s2.filegroups.f1 = /home/data/192.168.60.101/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f2 = /home/data/192.168.60.102/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f3 = /home/data/192.168.60.103/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f4 = /home/data/192.168.60.104/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f5 = /home/data/192.168.60.105/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f6 = /home/data/192.168.60.106/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f7 = /home/data/192.168.60.107/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f8 = /home/data/192.168.60.108/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f9 = /home/data/192.168.60.109/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f10 = /home/data/192.168.60.110/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f11 = /home/data/192.168.60.111/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f12 = /home/data/192.168.60.112/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+active_denfence.sources.s2.filegroups.f13 = /home/data/192.168.60.113/ACTIVE-DEFENCE-EVENT-LOG/.*[2-3].dat
+
+
+active_denfence.sources.s2.maxBatchCount = 10000
+
+# kafka channel充当生产者
+active_denfence.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
+active_denfence.channels.c2.brokerList = 192.168.1.3:9093
+active_denfence.channels.c2.topic = ACTIVE-DEFENCE-EVENT-LOG
+#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
+active_denfence.channels.c2.parseAsFlumeEvent = false
+active_denfence.channels.c2.kafka.producer.acks = 1
+active_denfence.channels.c2.producer.type=sync
+active_denfence.channels.c2.queue.buffering.max.ms = 5000
+active_denfence.channels.c2.queue.buffering.max.messages=20000
+active_denfence.channels.c2.batch.num.messages=5000
diff --git a/internal-flume/origin/conf/active/active_defence_f2k3.properties b/internal-flume/origin/conf/active/active_defence_f2k3.properties
new file mode 100644
index 0000000..69a9d26
--- /dev/null
+++ b/internal-flume/origin/conf/active/active_defence_f2k3.properties
@@ -0,0 +1,40 @@
+active_denfence.sources = s2
+active_denfence.channels = c2
+#指定source收集到的数据发送到哪个管道
+active_denfence.sources.s2.channels = c2
+#taildir source
+active_denfence.sources.s2.type = TAILDIR
+active_denfence.sources.s2.channels = c2
+#需要映射到宿主机做持久化
+active_denfence.sources.s2.positionFile = /home/test/3active_denfence_position.json
+#需要分组13台主机的数据
+active_denfence.sources.s2.filegroups = f1 f2 f3 f4 f5 f6 f7 f8 f9 f10 f11 f12 f13
+#需要分组按照序号 [0-9] 目前最多支持10个进程,可以增加
+active_denfence.sources.s2.filegroups.f1 = /home/data/192.168.60.101/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f2 = /home/data/192.168.60.102/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f3 = /home/data/192.168.60.103/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f4 = /home/data/192.168.60.104/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f5 = /home/data/192.168.60.105/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f6 = /home/data/192.168.60.106/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f7 = /home/data/192.168.60.107/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f8 = /home/data/192.168.60.108/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f9 = /home/data/192.168.60.109/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f10 = /home/data/192.168.60.110/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f11 = /home/data/192.168.60.111/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f12 = /home/data/192.168.60.112/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+active_denfence.sources.s2.filegroups.f13 = /home/data/192.168.60.113/ACTIVE-DEFENCE-EVENT-LOG/.*[4-5].dat
+
+
+active_denfence.sources.s2.maxBatchCount = 10000
+
+# kafka channel充当生产者
+active_denfence.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
+active_denfence.channels.c2.brokerList = 192.168.1.3:9093
+active_denfence.channels.c2.topic = ACTIVE-DEFENCE-EVENT-LOG
+#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
+active_denfence.channels.c2.parseAsFlumeEvent = false
+active_denfence.channels.c2.kafka.producer.acks = 1
+active_denfence.channels.c2.producer.type=sync
+active_denfence.channels.c2.queue.buffering.max.ms = 5000
+active_denfence.channels.c2.queue.buffering.max.messages=20000
+active_denfence.channels.c2.batch.num.messages=5000
diff --git a/internal-flume/origin/conf/active/active_defence_f2k4.properties b/internal-flume/origin/conf/active/active_defence_f2k4.properties
new file mode 100644
index 0000000..7037f00
--- /dev/null
+++ b/internal-flume/origin/conf/active/active_defence_f2k4.properties
@@ -0,0 +1,40 @@
+active_denfence.sources = s2
+active_denfence.channels = c2
+#指定source收集到的数据发送到哪个管道
+active_denfence.sources.s2.channels = c2
+#taildir source
+active_denfence.sources.s2.type = TAILDIR
+active_denfence.sources.s2.channels = c2
+#需要映射到宿主机做持久化
+active_denfence.sources.s2.positionFile = /home/test/4active_denfence_position.json
+#需要分组13台主机的数据
+active_denfence.sources.s2.filegroups = f1 f2 f3 f4 f5 f6 f7 f8 f9 f10 f11 f12 f13
+#需要分组按照序号 [0-9] 目前最多支持10个进程,可以增加
+active_denfence.sources.s2.filegroups.f1 = /home/data/192.168.60.101/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f2 = /home/data/192.168.60.102/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f3 = /home/data/192.168.60.103/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f4 = /home/data/192.168.60.104/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f5 = /home/data/192.168.60.105/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f6 = /home/data/192.168.60.106/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f7 = /home/data/192.168.60.107/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f8 = /home/data/192.168.60.108/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f9 = /home/data/192.168.60.109/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f10 = /home/data/192.168.60.110/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f11 = /home/data/192.168.60.111/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f12 = /home/data/192.168.60.112/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+active_denfence.sources.s2.filegroups.f13 = /home/data/192.168.60.113/ACTIVE-DEFENCE-EVENT-LOG/.*[6-7].dat
+
+
+active_denfence.sources.s2.maxBatchCount = 10000
+
+# kafka channel充当生产者
+active_denfence.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
+active_denfence.channels.c2.brokerList = 192.168.1.3:9093
+active_denfence.channels.c2.topic = ACTIVE-DEFENCE-EVENT-LOG
+#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
+active_denfence.channels.c2.parseAsFlumeEvent = false
+active_denfence.channels.c2.kafka.producer.acks = 1
+active_denfence.channels.c2.producer.type=sync
+active_denfence.channels.c2.queue.buffering.max.ms = 5000
+active_denfence.channels.c2.queue.buffering.max.messages=20000
+active_denfence.channels.c2.batch.num.messages=5000
diff --git a/internal-flume/origin/conf/active/active_defence_f2k5.properties b/internal-flume/origin/conf/active/active_defence_f2k5.properties
new file mode 100644
index 0000000..9b15583
--- /dev/null
+++ b/internal-flume/origin/conf/active/active_defence_f2k5.properties
@@ -0,0 +1,40 @@
+active_denfence.sources = s2
+active_denfence.channels = c2
+#指定source收集到的数据发送到哪个管道
+active_denfence.sources.s2.channels = c2
+#taildir source
+active_denfence.sources.s2.type = TAILDIR
+active_denfence.sources.s2.channels = c2
+#需要映射到宿主机做持久化
+active_denfence.sources.s2.positionFile = /home/test/5active_denfence_position.json
+#需要分组13台主机的数据
+active_denfence.sources.s2.filegroups = f1 f2 f3 f4 f5 f6 f7 f8 f9 f10 f11 f12 f13
+#需要分组按照序号 [0-9] 目前最多支持10个进程,可以增加
+active_denfence.sources.s2.filegroups.f1 = /home/data/192.168.60.101/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f2 = /home/data/192.168.60.102/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f3 = /home/data/192.168.60.103/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f4 = /home/data/192.168.60.104/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f5 = /home/data/192.168.60.105/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f6 = /home/data/192.168.60.106/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f7 = /home/data/192.168.60.107/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f8 = /home/data/192.168.60.108/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f9 = /home/data/192.168.60.109/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f10 = /home/data/192.168.60.110/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f11 = /home/data/192.168.60.111/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f12 = /home/data/192.168.60.112/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+active_denfence.sources.s2.filegroups.f13 = /home/data/192.168.60.113/ACTIVE-DEFENCE-EVENT-LOG/.*[8-9].dat
+
+
+active_denfence.sources.s2.maxBatchCount = 10000
+
+# kafka channel充当生产者
+active_denfence.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
+active_denfence.channels.c2.brokerList = 192.168.1.3:9093
+active_denfence.channels.c2.topic = ACTIVE-DEFENCE-EVENT-LOG
+#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
+active_denfence.channels.c2.parseAsFlumeEvent = false
+active_denfence.channels.c2.kafka.producer.acks = 1
+active_denfence.channels.c2.producer.type=sync
+active_denfence.channels.c2.queue.buffering.max.ms = 5000
+active_denfence.channels.c2.queue.buffering.max.messages=20000
+active_denfence.channels.c2.batch.num.messages=5000
diff --git a/internal-flume/conf/k2f/clean_start_flume.sh b/internal-flume/origin/conf/active/clean_start_flume.sh
index ef811a5..ef811a5 100755
--- a/internal-flume/conf/k2f/clean_start_flume.sh
+++ b/internal-flume/origin/conf/active/clean_start_flume.sh
diff --git a/internal-flume/conf/proxy/count_flume.sh b/internal-flume/origin/conf/active/count_flume.sh
index 4a50a2a..4a50a2a 100755
--- a/internal-flume/conf/proxy/count_flume.sh
+++ b/internal-flume/origin/conf/active/count_flume.sh
diff --git a/internal-flume/origin/conf/active/dae_active_denfence.sh b/internal-flume/origin/conf/active/dae_active_denfence.sh
new file mode 100755
index 0000000..8c2ed2e
--- /dev/null
+++ b/internal-flume/origin/conf/active/dae_active_denfence.sh
@@ -0,0 +1,39 @@
+#!/bin/sh
+
+#JAR_NAME=ktk_ip_asn.properties
+PROPERTIES_NAME=active_denfence_f2k
+#flume进程名称
+FLUME_NAME=active_denfence
+#flume根目录
+BASE_DIR=$(cd $(dirname $0); cd ../../; pwd)
+#当前路径
+CONF_DIR=$(cd $(dirname $0); pwd)
+#进程总数
+PROCESS_SUM=5
+
+echo "##############################################################" >> ${CONF_DIR}/restart_log/restart_${PROPERTIES_NAME}.log
+id=0 #配置文件后面的序号,无特殊作用
+flag=0 # flag为0表示初始化状态,为1表示完整启动成功所有进程,为2表示进程意外停止之后又重新杀死其余的进程并重新启动所有进程
+while true ; do
+ NUM=`ps -ef | grep ${PROPERTIES_NAME} | grep -v grep | grep -v dae |wc -l`
+ pids=$(ps -ef | grep ${PROPERTIES_NAME}\* | grep properties | awk '{print $2}')
+ time_stamp=$(date +%Y%m%d%H%M%S)
+ #大于设置进程数,杀掉所有进程,重启
+ if [[ "${NUM}" -ne ${PROCESS_SUM} && $flag -eq "1" ]];then
+ for pid in $pids
+ do
+ kill -9 $pid
+ done
+ flag=2
+ #如果正在运行的进程数小于定义的进程数,就启动
+ elif [ "${NUM}" -lt ${PROCESS_SUM} ];then
+ id=$(( ( ($id) % $PROCESS_SUM ) + 1 ))
+ nohup ${BASE_DIR}/bin/flume-ng agent -n ${FLUME_NAME} -c ${CONF_DIR} -f ${CONF_DIR}/${PROPERTIES_NAME}$id.properties >/dev/null 2>&1 &
+ echo "${time_stamp} ---> the ${PROPERTIES_NAME}_APP restart ---> $id" >> ${CONF_DIR}/restart_log/restart_${PROPERTIES_NAME}.log
+ #进程数如果等于指定的进程数,那就设置flag为1 id为0
+ elif [ "${NUM}" -eq ${PROCESS_SUM} ];then
+ flag=1
+ id=0
+ fi
+ sleep 10
+done
diff --git a/internal-flume/conf/k2f/flume-env.sh b/internal-flume/origin/conf/active/flume-env.sh
index 63979aa..63979aa 100755
--- a/internal-flume/conf/k2f/flume-env.sh
+++ b/internal-flume/origin/conf/active/flume-env.sh
diff --git a/internal-flume/origin/conf/active/restart_log/restart_connection_f2k.log b/internal-flume/origin/conf/active/restart_log/restart_connection_f2k.log
new file mode 100644
index 0000000..ce2b37d
--- /dev/null
+++ b/internal-flume/origin/conf/active/restart_log/restart_connection_f2k.log
@@ -0,0 +1,96 @@
+##############################################################
+20200512180513 ---> the connection_f2k_APP restart ---> 1
+20200512180523 ---> the connection_f2k_APP restart ---> 2
+20200512180533 ---> the connection_f2k_APP restart ---> 3
+##############################################################
+20200512180536 ---> the connection_f2k_APP restart ---> 1
+20200512180546 ---> the connection_f2k_APP restart ---> 2
+20200512180556 ---> the connection_f2k_APP restart ---> 3
+20200512180606 ---> the connection_f2k_APP restart ---> 4
+20200512180616 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200512183204 ---> the connection_f2k_APP restart ---> 1
+20200512183214 ---> the connection_f2k_APP restart ---> 2
+20200512183224 ---> the connection_f2k_APP restart ---> 3
+20200512183234 ---> the connection_f2k_APP restart ---> 4
+20200512183245 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200512184503 ---> the connection_f2k_APP restart ---> 1
+20200512184513 ---> the connection_f2k_APP restart ---> 2
+20200512184523 ---> the connection_f2k_APP restart ---> 3
+20200512184533 ---> the connection_f2k_APP restart ---> 4
+20200512184543 ---> the connection_f2k_APP restart ---> 5
+20200512191902 ---> the connection_f2k_APP restart ---> 1
+20200512191912 ---> the connection_f2k_APP restart ---> 2
+##############################################################
+20200512192032 ---> the connection_f2k_APP restart ---> 1
+20200512192043 ---> the connection_f2k_APP restart ---> 2
+20200512192053 ---> the connection_f2k_APP restart ---> 3
+20200512192103 ---> the connection_f2k_APP restart ---> 4
+20200512192113 ---> the connection_f2k_APP restart ---> 5
+20200512192323 ---> the connection_f2k_APP restart ---> 1
+20200512192333 ---> the connection_f2k_APP restart ---> 2
+20200512192343 ---> the connection_f2k_APP restart ---> 3
+20200512192353 ---> the connection_f2k_APP restart ---> 4
+20200512192403 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200512193146 ---> the connection_f2k_APP restart ---> 1
+20200512193156 ---> the connection_f2k_APP restart ---> 2
+20200512193206 ---> the connection_f2k_APP restart ---> 3
+20200512193216 ---> the connection_f2k_APP restart ---> 4
+20200512193226 ---> the connection_f2k_APP restart ---> 5
+20200512193406 ---> the connection_f2k_APP restart ---> 1
+20200512193416 ---> the connection_f2k_APP restart ---> 2
+20200512193426 ---> the connection_f2k_APP restart ---> 3
+20200512193437 ---> the connection_f2k_APP restart ---> 4
+##############################################################
+20200512193804 ---> the connection_f2k_APP restart ---> 1
+20200512193814 ---> the connection_f2k_APP restart ---> 2
+20200512193824 ---> the connection_f2k_APP restart ---> 3
+20200512193834 ---> the connection_f2k_APP restart ---> 4
+20200512193844 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200513102414 ---> the connection_f2k_APP restart ---> 1
+20200513102424 ---> the connection_f2k_APP restart ---> 2
+20200513102434 ---> the connection_f2k_APP restart ---> 3
+20200513102444 ---> the connection_f2k_APP restart ---> 4
+20200513102454 ---> the connection_f2k_APP restart ---> 5
+20200513104430 ---> the connection_f2k_APP restart ---> 1
+20200513104440 ---> the connection_f2k_APP restart ---> 2
+20200513104450 ---> the connection_f2k_APP restart ---> 3
+20200513104500 ---> the connection_f2k_APP restart ---> 4
+20200513104510 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200513181554 ---> the connection_f2k_APP restart ---> 1
+20200513181604 ---> the connection_f2k_APP restart ---> 2
+20200513181614 ---> the connection_f2k_APP restart ---> 3
+20200513181624 ---> the connection_f2k_APP restart ---> 4
+20200513181634 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200513182119 ---> the connection_f2k_APP restart ---> 1
+20200513182129 ---> the connection_f2k_APP restart ---> 2
+20200513182139 ---> the connection_f2k_APP restart ---> 3
+20200513182149 ---> the connection_f2k_APP restart ---> 4
+20200513182159 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200513182640 ---> the connection_f2k_APP restart ---> 1
+20200513182651 ---> the connection_f2k_APP restart ---> 2
+20200513182701 ---> the connection_f2k_APP restart ---> 3
+20200513182711 ---> the connection_f2k_APP restart ---> 4
+20200513182721 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200514171459 ---> the connection_f2k_APP restart ---> 1
+20200514171509 ---> the connection_f2k_APP restart ---> 2
+20200514171519 ---> the connection_f2k_APP restart ---> 3
+20200514171529 ---> the connection_f2k_APP restart ---> 4
+20200514171539 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200514172152 ---> the connection_f2k_APP restart ---> 1
+20200514172202 ---> the connection_f2k_APP restart ---> 2
+20200514172212 ---> the connection_f2k_APP restart ---> 3
+##############################################################
+20200514172234 ---> the connection_f2k_APP restart ---> 1
+20200514172244 ---> the connection_f2k_APP restart ---> 2
+20200514172254 ---> the connection_f2k_APP restart ---> 3
+20200514172304 ---> the connection_f2k_APP restart ---> 4
+20200514172314 ---> the connection_f2k_APP restart ---> 5
diff --git a/internal-flume/origin/conf/active/start_flume.sh b/internal-flume/origin/conf/active/start_flume.sh
new file mode 100755
index 0000000..fee77e9
--- /dev/null
+++ b/internal-flume/origin/conf/active/start_flume.sh
@@ -0,0 +1,4 @@
+#!/bin/sh
+
+BASE_DIR=$(cd $(dirname $0); pwd)
+nohup ${BASE_DIR}/dae_active_denfence.sh >/dev/null 2>&1 &
diff --git a/internal-flume/conf/proxy/stop_flume.sh b/internal-flume/origin/conf/active/stop_flume.sh
index 1d642b6..1d642b6 100755
--- a/internal-flume/conf/proxy/stop_flume.sh
+++ b/internal-flume/origin/conf/active/stop_flume.sh
diff --git a/internal-flume/conf/proxy/clean_start_flume.sh b/internal-flume/origin/conf/connection/clean_start_flume.sh
index ef811a5..ef811a5 100755
--- a/internal-flume/conf/proxy/clean_start_flume.sh
+++ b/internal-flume/origin/conf/connection/clean_start_flume.sh
diff --git a/internal-flume/conf/connection/connection_f2k1.properties b/internal-flume/origin/conf/connection/connection_f2k1.properties
index c61222e..11b449e 100644
--- a/internal-flume/conf/connection/connection_f2k1.properties
+++ b/internal-flume/origin/conf/connection/connection_f2k1.properties
@@ -29,7 +29,7 @@ connection.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
connection.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-connection.channels.c2.brokerList = 192.168.40.222:9093
+connection.channels.c2.brokerList = 192.168.1.3:9093
connection.channels.c2.topic = CONNECTION-RECORD-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
connection.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/connection/connection_f2k2.properties b/internal-flume/origin/conf/connection/connection_f2k2.properties
index 03baa50..7021daf 100644
--- a/internal-flume/conf/connection/connection_f2k2.properties
+++ b/internal-flume/origin/conf/connection/connection_f2k2.properties
@@ -29,7 +29,7 @@ connection.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
connection.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-connection.channels.c2.brokerList = 192.168.40.222:9093
+connection.channels.c2.brokerList = 192.168.1.3:9093
connection.channels.c2.topic = CONNECTION-RECORD-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
connection.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/connection/connection_f2k3.properties b/internal-flume/origin/conf/connection/connection_f2k3.properties
index cf472df..e634a22 100644
--- a/internal-flume/conf/connection/connection_f2k3.properties
+++ b/internal-flume/origin/conf/connection/connection_f2k3.properties
@@ -29,7 +29,7 @@ connection.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
connection.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-connection.channels.c2.brokerList = 192.168.40.222:9093
+connection.channels.c2.brokerList = 192.168.1.3:9093
connection.channels.c2.topic = CONNECTION-RECORD-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
connection.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/connection/connection_f2k4.properties b/internal-flume/origin/conf/connection/connection_f2k4.properties
index 9d109bb..d106aa3 100644
--- a/internal-flume/conf/connection/connection_f2k4.properties
+++ b/internal-flume/origin/conf/connection/connection_f2k4.properties
@@ -29,7 +29,7 @@ connection.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
connection.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-connection.channels.c2.brokerList = 192.168.40.222:9093
+connection.channels.c2.brokerList = 192.168.1.3:9093
connection.channels.c2.topic = CONNECTION-RECORD-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
connection.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/connection/connection_f2k5.properties b/internal-flume/origin/conf/connection/connection_f2k5.properties
index cd7c55d..6a02574 100644
--- a/internal-flume/conf/connection/connection_f2k5.properties
+++ b/internal-flume/origin/conf/connection/connection_f2k5.properties
@@ -29,7 +29,7 @@ connection.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
connection.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-connection.channels.c2.brokerList = 192.168.40.222:9093
+connection.channels.c2.brokerList = 192.168.1.3:9093
connection.channels.c2.topic = CONNECTION-RECORD-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
connection.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/radius/count_flume.sh b/internal-flume/origin/conf/connection/count_flume.sh
index 4a50a2a..4a50a2a 100755
--- a/internal-flume/conf/radius/count_flume.sh
+++ b/internal-flume/origin/conf/connection/count_flume.sh
diff --git a/internal-flume/conf/connection/dae_connection.sh b/internal-flume/origin/conf/connection/dae_connection.sh
index 3292604..3292604 100755
--- a/internal-flume/conf/connection/dae_connection.sh
+++ b/internal-flume/origin/conf/connection/dae_connection.sh
diff --git a/internal-flume/conf/proxy/flume-env.sh b/internal-flume/origin/conf/connection/flume-env.sh
index 63979aa..63979aa 100755
--- a/internal-flume/conf/proxy/flume-env.sh
+++ b/internal-flume/origin/conf/connection/flume-env.sh
diff --git a/internal-flume/origin/conf/connection/restart_log/restart_connection_f2k.log b/internal-flume/origin/conf/connection/restart_log/restart_connection_f2k.log
new file mode 100644
index 0000000..ce2b37d
--- /dev/null
+++ b/internal-flume/origin/conf/connection/restart_log/restart_connection_f2k.log
@@ -0,0 +1,96 @@
+##############################################################
+20200512180513 ---> the connection_f2k_APP restart ---> 1
+20200512180523 ---> the connection_f2k_APP restart ---> 2
+20200512180533 ---> the connection_f2k_APP restart ---> 3
+##############################################################
+20200512180536 ---> the connection_f2k_APP restart ---> 1
+20200512180546 ---> the connection_f2k_APP restart ---> 2
+20200512180556 ---> the connection_f2k_APP restart ---> 3
+20200512180606 ---> the connection_f2k_APP restart ---> 4
+20200512180616 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200512183204 ---> the connection_f2k_APP restart ---> 1
+20200512183214 ---> the connection_f2k_APP restart ---> 2
+20200512183224 ---> the connection_f2k_APP restart ---> 3
+20200512183234 ---> the connection_f2k_APP restart ---> 4
+20200512183245 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200512184503 ---> the connection_f2k_APP restart ---> 1
+20200512184513 ---> the connection_f2k_APP restart ---> 2
+20200512184523 ---> the connection_f2k_APP restart ---> 3
+20200512184533 ---> the connection_f2k_APP restart ---> 4
+20200512184543 ---> the connection_f2k_APP restart ---> 5
+20200512191902 ---> the connection_f2k_APP restart ---> 1
+20200512191912 ---> the connection_f2k_APP restart ---> 2
+##############################################################
+20200512192032 ---> the connection_f2k_APP restart ---> 1
+20200512192043 ---> the connection_f2k_APP restart ---> 2
+20200512192053 ---> the connection_f2k_APP restart ---> 3
+20200512192103 ---> the connection_f2k_APP restart ---> 4
+20200512192113 ---> the connection_f2k_APP restart ---> 5
+20200512192323 ---> the connection_f2k_APP restart ---> 1
+20200512192333 ---> the connection_f2k_APP restart ---> 2
+20200512192343 ---> the connection_f2k_APP restart ---> 3
+20200512192353 ---> the connection_f2k_APP restart ---> 4
+20200512192403 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200512193146 ---> the connection_f2k_APP restart ---> 1
+20200512193156 ---> the connection_f2k_APP restart ---> 2
+20200512193206 ---> the connection_f2k_APP restart ---> 3
+20200512193216 ---> the connection_f2k_APP restart ---> 4
+20200512193226 ---> the connection_f2k_APP restart ---> 5
+20200512193406 ---> the connection_f2k_APP restart ---> 1
+20200512193416 ---> the connection_f2k_APP restart ---> 2
+20200512193426 ---> the connection_f2k_APP restart ---> 3
+20200512193437 ---> the connection_f2k_APP restart ---> 4
+##############################################################
+20200512193804 ---> the connection_f2k_APP restart ---> 1
+20200512193814 ---> the connection_f2k_APP restart ---> 2
+20200512193824 ---> the connection_f2k_APP restart ---> 3
+20200512193834 ---> the connection_f2k_APP restart ---> 4
+20200512193844 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200513102414 ---> the connection_f2k_APP restart ---> 1
+20200513102424 ---> the connection_f2k_APP restart ---> 2
+20200513102434 ---> the connection_f2k_APP restart ---> 3
+20200513102444 ---> the connection_f2k_APP restart ---> 4
+20200513102454 ---> the connection_f2k_APP restart ---> 5
+20200513104430 ---> the connection_f2k_APP restart ---> 1
+20200513104440 ---> the connection_f2k_APP restart ---> 2
+20200513104450 ---> the connection_f2k_APP restart ---> 3
+20200513104500 ---> the connection_f2k_APP restart ---> 4
+20200513104510 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200513181554 ---> the connection_f2k_APP restart ---> 1
+20200513181604 ---> the connection_f2k_APP restart ---> 2
+20200513181614 ---> the connection_f2k_APP restart ---> 3
+20200513181624 ---> the connection_f2k_APP restart ---> 4
+20200513181634 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200513182119 ---> the connection_f2k_APP restart ---> 1
+20200513182129 ---> the connection_f2k_APP restart ---> 2
+20200513182139 ---> the connection_f2k_APP restart ---> 3
+20200513182149 ---> the connection_f2k_APP restart ---> 4
+20200513182159 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200513182640 ---> the connection_f2k_APP restart ---> 1
+20200513182651 ---> the connection_f2k_APP restart ---> 2
+20200513182701 ---> the connection_f2k_APP restart ---> 3
+20200513182711 ---> the connection_f2k_APP restart ---> 4
+20200513182721 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200514171459 ---> the connection_f2k_APP restart ---> 1
+20200514171509 ---> the connection_f2k_APP restart ---> 2
+20200514171519 ---> the connection_f2k_APP restart ---> 3
+20200514171529 ---> the connection_f2k_APP restart ---> 4
+20200514171539 ---> the connection_f2k_APP restart ---> 5
+##############################################################
+20200514172152 ---> the connection_f2k_APP restart ---> 1
+20200514172202 ---> the connection_f2k_APP restart ---> 2
+20200514172212 ---> the connection_f2k_APP restart ---> 3
+##############################################################
+20200514172234 ---> the connection_f2k_APP restart ---> 1
+20200514172244 ---> the connection_f2k_APP restart ---> 2
+20200514172254 ---> the connection_f2k_APP restart ---> 3
+20200514172304 ---> the connection_f2k_APP restart ---> 4
+20200514172314 ---> the connection_f2k_APP restart ---> 5
diff --git a/internal-flume/conf/connection/start_flume.sh b/internal-flume/origin/conf/connection/start_flume.sh
index e74045a..e74045a 100755
--- a/internal-flume/conf/connection/start_flume.sh
+++ b/internal-flume/origin/conf/connection/start_flume.sh
diff --git a/internal-flume/conf/radius/stop_flume.sh b/internal-flume/origin/conf/connection/stop_flume.sh
index 1d642b6..1d642b6 100755
--- a/internal-flume/conf/radius/stop_flume.sh
+++ b/internal-flume/origin/conf/connection/stop_flume.sh
diff --git a/internal-flume/conf/radius/clean_start_flume.sh b/internal-flume/origin/conf/proxy/clean_start_flume.sh
index ef811a5..ef811a5 100755
--- a/internal-flume/conf/radius/clean_start_flume.sh
+++ b/internal-flume/origin/conf/proxy/clean_start_flume.sh
diff --git a/internal-flume/conf/security/count_flume.sh b/internal-flume/origin/conf/proxy/count_flume.sh
index 4a50a2a..4a50a2a 100755
--- a/internal-flume/conf/security/count_flume.sh
+++ b/internal-flume/origin/conf/proxy/count_flume.sh
diff --git a/internal-flume/conf/proxy/dae_proxy.sh b/internal-flume/origin/conf/proxy/dae_proxy.sh
index d9746eb..d9746eb 100755
--- a/internal-flume/conf/proxy/dae_proxy.sh
+++ b/internal-flume/origin/conf/proxy/dae_proxy.sh
diff --git a/internal-flume/conf/radius/flume-env.sh b/internal-flume/origin/conf/proxy/flume-env.sh
index 63979aa..63979aa 100755
--- a/internal-flume/conf/radius/flume-env.sh
+++ b/internal-flume/origin/conf/proxy/flume-env.sh
diff --git a/internal-flume/conf/proxy/proxy_f2k1.properties b/internal-flume/origin/conf/proxy/proxy_f2k1.properties
index 7a7297b..ca9eec3 100644
--- a/internal-flume/conf/proxy/proxy_f2k1.properties
+++ b/internal-flume/origin/conf/proxy/proxy_f2k1.properties
@@ -29,7 +29,7 @@ proxy.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
proxy.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-proxy.channels.c2.brokerList = 192.168.40.222:9093
+proxy.channels.c2.brokerList = 192.168.1.3:9093
proxy.channels.c2.topic = PROXY-EVENT-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
proxy.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/proxy/proxy_f2k2.properties b/internal-flume/origin/conf/proxy/proxy_f2k2.properties
index a1ff401..e90b4f4 100644
--- a/internal-flume/conf/proxy/proxy_f2k2.properties
+++ b/internal-flume/origin/conf/proxy/proxy_f2k2.properties
@@ -29,7 +29,7 @@ proxy.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
proxy.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-proxy.channels.c2.brokerList = 192.168.40.222:9093
+proxy.channels.c2.brokerList = 192.168.1.3:9093
proxy.channels.c2.topic = PROXY-EVENT-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
proxy.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/proxy/proxy_f2k3.properties b/internal-flume/origin/conf/proxy/proxy_f2k3.properties
index e06b5b6..40831f1 100644
--- a/internal-flume/conf/proxy/proxy_f2k3.properties
+++ b/internal-flume/origin/conf/proxy/proxy_f2k3.properties
@@ -29,7 +29,7 @@ proxy.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
proxy.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-proxy.channels.c2.brokerList = 192.168.40.222:9093
+proxy.channels.c2.brokerList = 192.168.1.3:9093
proxy.channels.c2.topic = PROXY-EVENT-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
proxy.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/proxy/proxy_f2k4.properties b/internal-flume/origin/conf/proxy/proxy_f2k4.properties
index 071c2d3..037e0da 100644
--- a/internal-flume/conf/proxy/proxy_f2k4.properties
+++ b/internal-flume/origin/conf/proxy/proxy_f2k4.properties
@@ -29,7 +29,7 @@ proxy.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
proxy.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-proxy.channels.c2.brokerList = 192.168.40.222:9093
+proxy.channels.c2.brokerList = 192.168.1.3:9093
proxy.channels.c2.topic = PROXY-EVENT-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
proxy.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/proxy/proxy_f2k5.properties b/internal-flume/origin/conf/proxy/proxy_f2k5.properties
index 4c08edb..3d046a9 100644
--- a/internal-flume/conf/proxy/proxy_f2k5.properties
+++ b/internal-flume/origin/conf/proxy/proxy_f2k5.properties
@@ -29,7 +29,7 @@ proxy.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
proxy.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-proxy.channels.c2.brokerList = 192.168.40.222:9093
+proxy.channels.c2.brokerList = 192.168.1.3:9093
proxy.channels.c2.topic = PROXY-EVENT-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
proxy.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/origin/conf/proxy/restart_log/restart_proxy_f2k.log b/internal-flume/origin/conf/proxy/restart_log/restart_proxy_f2k.log
new file mode 100644
index 0000000..e0f112c
--- /dev/null
+++ b/internal-flume/origin/conf/proxy/restart_log/restart_proxy_f2k.log
@@ -0,0 +1,74 @@
+##############################################################
+20200512174751 ---> the proxy_f2k_APP restart ---> 1
+20200512174801 ---> the proxy_f2k_APP restart ---> 2
+20200512174811 ---> the proxy_f2k_APP restart ---> 3
+20200512174821 ---> the proxy_f2k_APP restart ---> 4
+20200512174831 ---> the proxy_f2k_APP restart ---> 5
+##############################################################
+20200512175511 ---> the proxy_f2k_APP restart ---> 1
+20200512175521 ---> the proxy_f2k_APP restart ---> 2
+20200512175531 ---> the proxy_f2k_APP restart ---> 3
+20200512175541 ---> the proxy_f2k_APP restart ---> 4
+20200512175551 ---> the proxy_f2k_APP restart ---> 5
+##############################################################
+20200512180513 ---> the proxy_f2k_APP restart ---> 1
+20200512180523 ---> the proxy_f2k_APP restart ---> 2
+20200512180533 ---> the proxy_f2k_APP restart ---> 3
+##############################################################
+20200512180536 ---> the proxy_f2k_APP restart ---> 1
+20200512180546 ---> the proxy_f2k_APP restart ---> 2
+20200512180556 ---> the proxy_f2k_APP restart ---> 3
+20200512180606 ---> the proxy_f2k_APP restart ---> 4
+20200512180616 ---> the proxy_f2k_APP restart ---> 5
+##############################################################
+20200512183204 ---> the proxy_f2k_APP restart ---> 1
+20200512183214 ---> the proxy_f2k_APP restart ---> 2
+20200512183224 ---> the proxy_f2k_APP restart ---> 3
+20200512183234 ---> the proxy_f2k_APP restart ---> 4
+20200512183245 ---> the proxy_f2k_APP restart ---> 5
+##############################################################
+20200512184503 ---> the proxy_f2k_APP restart ---> 1
+20200512184513 ---> the proxy_f2k_APP restart ---> 2
+20200512184523 ---> the proxy_f2k_APP restart ---> 3
+20200512184533 ---> the proxy_f2k_APP restart ---> 4
+20200512184543 ---> the proxy_f2k_APP restart ---> 5
+##############################################################
+20200513102414 ---> the proxy_f2k_APP restart ---> 1
+20200513102424 ---> the proxy_f2k_APP restart ---> 2
+20200513102434 ---> the proxy_f2k_APP restart ---> 3
+20200513102444 ---> the proxy_f2k_APP restart ---> 4
+20200513102454 ---> the proxy_f2k_APP restart ---> 5
+##############################################################
+20200513181554 ---> the proxy_f2k_APP restart ---> 1
+20200513181604 ---> the proxy_f2k_APP restart ---> 2
+20200513181614 ---> the proxy_f2k_APP restart ---> 3
+20200513181624 ---> the proxy_f2k_APP restart ---> 4
+20200513181634 ---> the proxy_f2k_APP restart ---> 5
+##############################################################
+20200513182119 ---> the proxy_f2k_APP restart ---> 1
+20200513182129 ---> the proxy_f2k_APP restart ---> 2
+20200513182139 ---> the proxy_f2k_APP restart ---> 3
+20200513182149 ---> the proxy_f2k_APP restart ---> 4
+20200513182159 ---> the proxy_f2k_APP restart ---> 5
+##############################################################
+20200513182640 ---> the proxy_f2k_APP restart ---> 1
+20200513182651 ---> the proxy_f2k_APP restart ---> 2
+20200513182701 ---> the proxy_f2k_APP restart ---> 3
+20200513182711 ---> the proxy_f2k_APP restart ---> 4
+20200513182721 ---> the proxy_f2k_APP restart ---> 5
+##############################################################
+20200514171459 ---> the proxy_f2k_APP restart ---> 1
+20200514171509 ---> the proxy_f2k_APP restart ---> 2
+20200514171519 ---> the proxy_f2k_APP restart ---> 3
+20200514171529 ---> the proxy_f2k_APP restart ---> 4
+20200514171539 ---> the proxy_f2k_APP restart ---> 5
+##############################################################
+20200514172152 ---> the proxy_f2k_APP restart ---> 1
+20200514172202 ---> the proxy_f2k_APP restart ---> 2
+20200514172212 ---> the proxy_f2k_APP restart ---> 3
+##############################################################
+20200514172234 ---> the proxy_f2k_APP restart ---> 1
+20200514172244 ---> the proxy_f2k_APP restart ---> 2
+20200514172254 ---> the proxy_f2k_APP restart ---> 3
+20200514172304 ---> the proxy_f2k_APP restart ---> 4
+20200514172314 ---> the proxy_f2k_APP restart ---> 5
diff --git a/internal-flume/origin/conf/proxy/start_flume.sh b/internal-flume/origin/conf/proxy/start_flume.sh
new file mode 100755
index 0000000..eb3662f
--- /dev/null
+++ b/internal-flume/origin/conf/proxy/start_flume.sh
@@ -0,0 +1,5 @@
+#!/bin/sh
+
+BASE_DIR=$(cd $(dirname $0); pwd)
+
+nohup ${BASE_DIR}/dae_proxy.sh >/dev/null 2>&1 &
diff --git a/internal-flume/conf/security/stop_flume.sh b/internal-flume/origin/conf/proxy/stop_flume.sh
index 1d642b6..1d642b6 100755
--- a/internal-flume/conf/security/stop_flume.sh
+++ b/internal-flume/origin/conf/proxy/stop_flume.sh
diff --git a/internal-flume/conf/security/clean_start_flume.sh b/internal-flume/origin/conf/radius/clean_start_flume.sh
index ef811a5..ef811a5 100755
--- a/internal-flume/conf/security/clean_start_flume.sh
+++ b/internal-flume/origin/conf/radius/clean_start_flume.sh
diff --git a/internal-flume/conf/k2f/count_flume.sh b/internal-flume/origin/conf/radius/count_flume.sh
index cc8452b..4a50a2a 100755
--- a/internal-flume/conf/k2f/count_flume.sh
+++ b/internal-flume/origin/conf/radius/count_flume.sh
@@ -2,12 +2,12 @@
BASE_DIR=$(cd $(dirname $0); pwd)
-JAR_NAME=`ls ${BASE_DIR} | grep ^k2* | grep .properties$`
+JAR_NAME=`ls ${BASE_DIR} | grep f2k* | grep .properties$`
NUM1=`ps -ef | grep ${JAR_NAME} | grep -v grep | wc -l`
pids1=$(ps -ef | grep ${JAR_NAME} | grep -v grep | awk '{print $2}')
echo 'flume '${JAR_NAME}' total process-->'${NUM1}
-if [ "${NUM1}" -ge "1" ];then
+if [ "${NUM1}" -ge "5" ];then
for pid1 in $pids1
do
echo 'flume '${JAR_NAME}' process-->'$pid1
diff --git a/internal-flume/conf/radius/dae_radius.sh b/internal-flume/origin/conf/radius/dae_radius.sh
index de5269d..de5269d 100755
--- a/internal-flume/conf/radius/dae_radius.sh
+++ b/internal-flume/origin/conf/radius/dae_radius.sh
diff --git a/internal-flume/conf/security/flume-env.sh b/internal-flume/origin/conf/radius/flume-env.sh
index 63979aa..63979aa 100755
--- a/internal-flume/conf/security/flume-env.sh
+++ b/internal-flume/origin/conf/radius/flume-env.sh
diff --git a/internal-flume/conf/radius/radius_f2k1.properties b/internal-flume/origin/conf/radius/radius_f2k1.properties
index 0029ac8..e30c49e 100644
--- a/internal-flume/conf/radius/radius_f2k1.properties
+++ b/internal-flume/origin/conf/radius/radius_f2k1.properties
@@ -29,7 +29,7 @@ radius.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
radius.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-radius.channels.c2.brokerList = 192.168.40.222:9093
+radius.channels.c2.brokerList = 192.168.1.3:9093
radius.channels.c2.topic = RADIUS-RECORD-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
radius.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/radius/radius_f2k2.properties b/internal-flume/origin/conf/radius/radius_f2k2.properties
index ccec9bc..68226ce 100644
--- a/internal-flume/conf/radius/radius_f2k2.properties
+++ b/internal-flume/origin/conf/radius/radius_f2k2.properties
@@ -29,7 +29,7 @@ radius.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
radius.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-radius.channels.c2.brokerList = 192.168.40.222:9093
+radius.channels.c2.brokerList = 192.168.1.3:9093
radius.channels.c2.topic = RADIUS-RECORD-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
radius.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/radius/radius_f2k3.properties b/internal-flume/origin/conf/radius/radius_f2k3.properties
index ef40bde..3ff051e 100644
--- a/internal-flume/conf/radius/radius_f2k3.properties
+++ b/internal-flume/origin/conf/radius/radius_f2k3.properties
@@ -29,7 +29,7 @@ radius.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
radius.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-radius.channels.c2.brokerList = 192.168.40.222:9093
+radius.channels.c2.brokerList = 192.168.1.3:9093
radius.channels.c2.topic = RADIUS-RECORD-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
radius.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/radius/radius_f2k4.properties b/internal-flume/origin/conf/radius/radius_f2k4.properties
index c62c3d6..cd3b1bf 100644
--- a/internal-flume/conf/radius/radius_f2k4.properties
+++ b/internal-flume/origin/conf/radius/radius_f2k4.properties
@@ -29,7 +29,7 @@ radius.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
radius.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-radius.channels.c2.brokerList = 192.168.40.222:9093
+radius.channels.c2.brokerList = 192.168.1.3:9093
radius.channels.c2.topic = RADIUS-RECORD-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
radius.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/radius/radius_f2k5.properties b/internal-flume/origin/conf/radius/radius_f2k5.properties
index 5c72446..a6d0ac3 100644
--- a/internal-flume/conf/radius/radius_f2k5.properties
+++ b/internal-flume/origin/conf/radius/radius_f2k5.properties
@@ -29,7 +29,7 @@ radius.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
radius.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-radius.channels.c2.brokerList = 192.168.40.222:9093
+radius.channels.c2.brokerList = 192.168.1.3:9093
radius.channels.c2.topic = RADIUS-RECORD-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
radius.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/origin/conf/radius/restart_log/restart_radius_f2k.log b/internal-flume/origin/conf/radius/restart_log/restart_radius_f2k.log
new file mode 100644
index 0000000..2d71ffc
--- /dev/null
+++ b/internal-flume/origin/conf/radius/restart_log/restart_radius_f2k.log
@@ -0,0 +1,68 @@
+##############################################################
+20200512173126 ---> the radius_f2k_APP restart ---> 1
+20200512173136 ---> the radius_f2k_APP restart ---> 2
+20200512173146 ---> the radius_f2k_APP restart ---> 3
+20200512173156 ---> the radius_f2k_APP restart ---> 4
+20200512173206 ---> the radius_f2k_APP restart ---> 5
+##############################################################
+20200512180513 ---> the radius_f2k_APP restart ---> 1
+20200512180523 ---> the radius_f2k_APP restart ---> 2
+20200512180533 ---> the radius_f2k_APP restart ---> 3
+##############################################################
+20200512180536 ---> the radius_f2k_APP restart ---> 1
+20200512180546 ---> the radius_f2k_APP restart ---> 2
+20200512180556 ---> the radius_f2k_APP restart ---> 3
+20200512180606 ---> the radius_f2k_APP restart ---> 4
+20200512180616 ---> the radius_f2k_APP restart ---> 5
+##############################################################
+20200512183204 ---> the radius_f2k_APP restart ---> 1
+20200512183214 ---> the radius_f2k_APP restart ---> 2
+20200512183224 ---> the radius_f2k_APP restart ---> 3
+20200512183235 ---> the radius_f2k_APP restart ---> 4
+20200512183245 ---> the radius_f2k_APP restart ---> 5
+##############################################################
+20200512184503 ---> the radius_f2k_APP restart ---> 1
+20200512184513 ---> the radius_f2k_APP restart ---> 2
+20200512184523 ---> the radius_f2k_APP restart ---> 3
+20200512184533 ---> the radius_f2k_APP restart ---> 4
+20200512184543 ---> the radius_f2k_APP restart ---> 5
+##############################################################
+20200513102414 ---> the radius_f2k_APP restart ---> 1
+20200513102424 ---> the radius_f2k_APP restart ---> 2
+20200513102434 ---> the radius_f2k_APP restart ---> 3
+20200513102444 ---> the radius_f2k_APP restart ---> 4
+20200513102455 ---> the radius_f2k_APP restart ---> 5
+##############################################################
+20200513181554 ---> the radius_f2k_APP restart ---> 1
+20200513181604 ---> the radius_f2k_APP restart ---> 2
+20200513181614 ---> the radius_f2k_APP restart ---> 3
+20200513181624 ---> the radius_f2k_APP restart ---> 4
+20200513181634 ---> the radius_f2k_APP restart ---> 5
+##############################################################
+20200513182119 ---> the radius_f2k_APP restart ---> 1
+20200513182129 ---> the radius_f2k_APP restart ---> 2
+20200513182139 ---> the radius_f2k_APP restart ---> 3
+20200513182149 ---> the radius_f2k_APP restart ---> 4
+20200513182159 ---> the radius_f2k_APP restart ---> 5
+##############################################################
+20200513182641 ---> the radius_f2k_APP restart ---> 1
+20200513182651 ---> the radius_f2k_APP restart ---> 2
+20200513182701 ---> the radius_f2k_APP restart ---> 3
+20200513182711 ---> the radius_f2k_APP restart ---> 4
+20200513182721 ---> the radius_f2k_APP restart ---> 5
+##############################################################
+20200514171459 ---> the radius_f2k_APP restart ---> 1
+20200514171509 ---> the radius_f2k_APP restart ---> 2
+20200514171519 ---> the radius_f2k_APP restart ---> 3
+20200514171529 ---> the radius_f2k_APP restart ---> 4
+20200514171539 ---> the radius_f2k_APP restart ---> 5
+##############################################################
+20200514172152 ---> the radius_f2k_APP restart ---> 1
+20200514172202 ---> the radius_f2k_APP restart ---> 2
+20200514172213 ---> the radius_f2k_APP restart ---> 3
+##############################################################
+20200514172234 ---> the radius_f2k_APP restart ---> 1
+20200514172244 ---> the radius_f2k_APP restart ---> 2
+20200514172254 ---> the radius_f2k_APP restart ---> 3
+20200514172304 ---> the radius_f2k_APP restart ---> 4
+20200514172314 ---> the radius_f2k_APP restart ---> 5
diff --git a/internal-flume/origin/conf/radius/start_flume.sh b/internal-flume/origin/conf/radius/start_flume.sh
new file mode 100755
index 0000000..8a4be32
--- /dev/null
+++ b/internal-flume/origin/conf/radius/start_flume.sh
@@ -0,0 +1,5 @@
+#!/bin/sh
+
+BASE_DIR=$(cd $(dirname $0); pwd)
+
+nohup ${BASE_DIR}/dae_radius.sh >/dev/null 2>&1 &
diff --git a/internal-flume/conf/k2f/stop_flume.sh b/internal-flume/origin/conf/radius/stop_flume.sh
index e4417aa..1d642b6 100755
--- a/internal-flume/conf/k2f/stop_flume.sh
+++ b/internal-flume/origin/conf/radius/stop_flume.sh
@@ -5,7 +5,7 @@
BASE_DIR=$(cd $(dirname $0); pwd)
DAE_NAME=`ls ${BASE_DIR} | grep ^dae_* | grep .sh$`
-JAR_NAME=`ls ${BASE_DIR} | grep ^f2* | grep .properties$`
+JAR_NAME=`ls ${BASE_DIR} | grep f2k\* | grep .properties$`
NUM1=`ps -ef | grep ${DAE_NAME} | grep -v grep | wc -l`
pids1=$(ps -ef | grep ${DAE_NAME} | grep -v grep | awk '{print $2}')
@@ -17,8 +17,10 @@ if [ "${NUM1}" -ge "1" ];then
done
fi
-NUM2=`ps -ef | grep ${JAR_NAME} | grep -v grep | wc -l`
-pids2=$(ps -ef | grep ${JAR_NAME} | grep -v grep | awk '{print $2}')
+for jar in $JAR_NAME ; do
+
+NUM2=`ps -ef | grep $jar | grep -v grep | wc -l`
+pids2=$(ps -ef | grep $jar | grep -v grep | awk '{print $2}')
if [ "${NUM2}" -ge "1" ];then
for pid2 in $pids2
do
@@ -26,3 +28,4 @@ if [ "${NUM2}" -ge "1" ];then
echo 'killed '${JAR_NAME}' process-->'$pid2
done
fi
+done
diff --git a/internal-flume/origin/conf/security/clean_start_flume.sh b/internal-flume/origin/conf/security/clean_start_flume.sh
new file mode 100755
index 0000000..ef811a5
--- /dev/null
+++ b/internal-flume/origin/conf/security/clean_start_flume.sh
@@ -0,0 +1,10 @@
+#!/bin/sh
+
+BASE_DIR=$(cd $(dirname $0); pwd)
+DAE_NAME=`ls ${BASE_DIR} | grep ^dae_* | grep .sh$`
+
+
+#清除日志和checkpoint以及历史data
+rm -rf ${BASE_DIR}/logs/*
+
+nohup ${BASE_DIR}/${DAE_NAME} $1 $2 >/dev/null 2>&1 &
diff --git a/internal-flume/origin/conf/security/count_flume.sh b/internal-flume/origin/conf/security/count_flume.sh
new file mode 100755
index 0000000..4a50a2a
--- /dev/null
+++ b/internal-flume/origin/conf/security/count_flume.sh
@@ -0,0 +1,15 @@
+#!/bin/sh
+
+BASE_DIR=$(cd $(dirname $0); pwd)
+
+JAR_NAME=`ls ${BASE_DIR} | grep f2k* | grep .properties$`
+
+NUM1=`ps -ef | grep ${JAR_NAME} | grep -v grep | wc -l`
+pids1=$(ps -ef | grep ${JAR_NAME} | grep -v grep | awk '{print $2}')
+echo 'flume '${JAR_NAME}' total process-->'${NUM1}
+if [ "${NUM1}" -ge "5" ];then
+ for pid1 in $pids1
+ do
+ echo 'flume '${JAR_NAME}' process-->'$pid1
+ done
+fi
diff --git a/internal-flume/conf/security/dae_security.sh b/internal-flume/origin/conf/security/dae_security.sh
index 54d5fa4..54d5fa4 100755
--- a/internal-flume/conf/security/dae_security.sh
+++ b/internal-flume/origin/conf/security/dae_security.sh
diff --git a/internal-flume/origin/conf/security/flume-env.sh b/internal-flume/origin/conf/security/flume-env.sh
new file mode 100755
index 0000000..63979aa
--- /dev/null
+++ b/internal-flume/origin/conf/security/flume-env.sh
@@ -0,0 +1,35 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements. See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership. The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License. You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+# If this file is placed at FLUME_CONF_DIR/flume-env.sh, it will be sourced
+# during Flume startup.
+
+# Enviroment variables can be set here.
+
+export JAVA_HOME=/usr/lib/jvm/jdk1.8.0_73
+
+# Give Flume more memory and pre-allocate, enable remote monitoring via JMX
+#export JAVA_OPTS="-Xms1024m -Xmx3072m -Dcom.sun.management.jmxremote"
+export JAVA_OPTS="-Xms512m -Xmx2048m -Xss256k -Xmn1g -XX:+UseParNewGC -XX:+UseConcMarkSweepGC -XX:-UseGCOverheadLimit -Dcom.sun.management.jmxremote"
+
+# Let Flume write raw event data and configuration information to its log files for debugging
+# purposes. Enabling these flags is not recommended in production,
+# as it may result in logging sensitive user information or encryption secrets.
+# export JAVA_OPTS="$JAVA_OPTS -Dorg.apache.flume.log.rawdata=true -Dorg.apache.flume.log.printconfig=true "
+
+# Note that the Flume conf directory is always included in the classpath.
+#FLUME_CLASSPATH=""
+
diff --git a/internal-flume/conf/security/restart_log/restart_security_f2k.log b/internal-flume/origin/conf/security/restart_log/restart_security_f2k.log
index 6ab01de..713a7bd 100644
--- a/internal-flume/conf/security/restart_log/restart_security_f2k.log
+++ b/internal-flume/origin/conf/security/restart_log/restart_security_f2k.log
@@ -149,3 +149,48 @@
20200512184523 ---> the security_f2k_APP restart ---> 3
20200512184533 ---> the security_f2k_APP restart ---> 4
20200512184543 ---> the security_f2k_APP restart ---> 5
+##############################################################
+20200513102414 ---> the security_f2k_APP restart ---> 1
+20200513102424 ---> the security_f2k_APP restart ---> 2
+20200513102434 ---> the security_f2k_APP restart ---> 3
+20200513102444 ---> the security_f2k_APP restart ---> 4
+20200513102455 ---> the security_f2k_APP restart ---> 5
+20200513104430 ---> the security_f2k_APP restart ---> 1
+20200513104440 ---> the security_f2k_APP restart ---> 2
+20200513104450 ---> the security_f2k_APP restart ---> 3
+20200513104500 ---> the security_f2k_APP restart ---> 4
+20200513104510 ---> the security_f2k_APP restart ---> 5
+##############################################################
+20200513181554 ---> the security_f2k_APP restart ---> 1
+20200513181604 ---> the security_f2k_APP restart ---> 2
+20200513181614 ---> the security_f2k_APP restart ---> 3
+20200513181624 ---> the security_f2k_APP restart ---> 4
+20200513181634 ---> the security_f2k_APP restart ---> 5
+##############################################################
+20200513182119 ---> the security_f2k_APP restart ---> 1
+20200513182129 ---> the security_f2k_APP restart ---> 2
+20200513182139 ---> the security_f2k_APP restart ---> 3
+20200513182149 ---> the security_f2k_APP restart ---> 4
+20200513182159 ---> the security_f2k_APP restart ---> 5
+##############################################################
+20200513182641 ---> the security_f2k_APP restart ---> 1
+20200513182651 ---> the security_f2k_APP restart ---> 2
+20200513182701 ---> the security_f2k_APP restart ---> 3
+20200513182711 ---> the security_f2k_APP restart ---> 4
+20200513182721 ---> the security_f2k_APP restart ---> 5
+##############################################################
+20200514171459 ---> the security_f2k_APP restart ---> 1
+20200514171509 ---> the security_f2k_APP restart ---> 2
+20200514171519 ---> the security_f2k_APP restart ---> 3
+20200514171529 ---> the security_f2k_APP restart ---> 4
+20200514171539 ---> the security_f2k_APP restart ---> 5
+##############################################################
+20200514172152 ---> the security_f2k_APP restart ---> 1
+20200514172203 ---> the security_f2k_APP restart ---> 2
+20200514172213 ---> the security_f2k_APP restart ---> 3
+##############################################################
+20200514172234 ---> the security_f2k_APP restart ---> 1
+20200514172244 ---> the security_f2k_APP restart ---> 2
+20200514172254 ---> the security_f2k_APP restart ---> 3
+20200514172304 ---> the security_f2k_APP restart ---> 4
+20200514172314 ---> the security_f2k_APP restart ---> 5
diff --git a/internal-flume/conf/security/security_f2k1.properties b/internal-flume/origin/conf/security/security_f2k1.properties
index 84db2f1..2a639e8 100644
--- a/internal-flume/conf/security/security_f2k1.properties
+++ b/internal-flume/origin/conf/security/security_f2k1.properties
@@ -29,7 +29,7 @@ security.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
security.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-security.channels.c2.brokerList = 192.168.40.222:9093
+security.channels.c2.brokerList = 192.168.1.3:9093
security.channels.c2.topic = SECURITY-EVENT-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
security.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/security/security_f2k2.properties b/internal-flume/origin/conf/security/security_f2k2.properties
index cbe2553..5a66271 100644
--- a/internal-flume/conf/security/security_f2k2.properties
+++ b/internal-flume/origin/conf/security/security_f2k2.properties
@@ -29,7 +29,7 @@ security.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
security.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-security.channels.c2.brokerList = 192.168.40.222:9093
+security.channels.c2.brokerList = 192.168.1.3:9093
security.channels.c2.topic = SECURITY-EVENT-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
security.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/security/security_f2k3.properties b/internal-flume/origin/conf/security/security_f2k3.properties
index 5399e40..1f8615d 100644
--- a/internal-flume/conf/security/security_f2k3.properties
+++ b/internal-flume/origin/conf/security/security_f2k3.properties
@@ -29,7 +29,7 @@ security.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
security.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-security.channels.c2.brokerList = 192.168.40.222:9093
+security.channels.c2.brokerList = 192.168.1.3:9093
security.channels.c2.topic = SECURITY-EVENT-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
security.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/security/security_f2k4.properties b/internal-flume/origin/conf/security/security_f2k4.properties
index 9933dad..fe71097 100644
--- a/internal-flume/conf/security/security_f2k4.properties
+++ b/internal-flume/origin/conf/security/security_f2k4.properties
@@ -29,7 +29,7 @@ security.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
security.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-security.channels.c2.brokerList = 192.168.40.222:9093
+security.channels.c2.brokerList = 192.168.1.3:9093
security.channels.c2.topic = SECURITY-EVENT-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
security.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/conf/security/security_f2k5.properties b/internal-flume/origin/conf/security/security_f2k5.properties
index 9fddcca..0e60059 100644
--- a/internal-flume/conf/security/security_f2k5.properties
+++ b/internal-flume/origin/conf/security/security_f2k5.properties
@@ -29,7 +29,7 @@ security.sources.s2.maxBatchCount = 10000
# kafka channel充当生产者
security.channels.c2.type = org.apache.flume.channel.kafka.KafkaChannel
-security.channels.c2.brokerList = 192.168.40.222:9093
+security.channels.c2.brokerList = 192.168.1.3:9093
security.channels.c2.topic = SECURITY-EVENT-LOG
#false表示是以纯文本的形式写进入的,true是以event的形式写进入的,以event写进入时,会出现乱码, 默认是true
security.channels.c2.parseAsFlumeEvent = false
diff --git a/internal-flume/origin/conf/security/start_flume.sh b/internal-flume/origin/conf/security/start_flume.sh
new file mode 100755
index 0000000..4e6eb06
--- /dev/null
+++ b/internal-flume/origin/conf/security/start_flume.sh
@@ -0,0 +1,5 @@
+#!/bin/sh
+
+BASE_DIR=$(cd $(dirname $0); pwd)
+
+nohup ${BASE_DIR}/dae_security.sh >/dev/null 2>&1 &
diff --git a/internal-flume/origin/conf/security/stop_flume.sh b/internal-flume/origin/conf/security/stop_flume.sh
new file mode 100755
index 0000000..1d642b6
--- /dev/null
+++ b/internal-flume/origin/conf/security/stop_flume.sh
@@ -0,0 +1,31 @@
+#!/bin/sh
+
+#DAE_NAME=dae_k2ha.sh
+#JAR_NAME=k2ha.properties
+BASE_DIR=$(cd $(dirname $0); pwd)
+
+DAE_NAME=`ls ${BASE_DIR} | grep ^dae_* | grep .sh$`
+JAR_NAME=`ls ${BASE_DIR} | grep f2k\* | grep .properties$`
+
+NUM1=`ps -ef | grep ${DAE_NAME} | grep -v grep | wc -l`
+pids1=$(ps -ef | grep ${DAE_NAME} | grep -v grep | awk '{print $2}')
+if [ "${NUM1}" -ge "1" ];then
+ for pid1 in $pids1
+ do
+ kill -9 $pid1
+ echo 'killed '${DAE_NAME}' process-->'$pid1
+ done
+fi
+
+for jar in $JAR_NAME ; do
+
+NUM2=`ps -ef | grep $jar | grep -v grep | wc -l`
+pids2=$(ps -ef | grep $jar | grep -v grep | awk '{print $2}')
+if [ "${NUM2}" -ge "1" ];then
+ for pid2 in $pids2
+ do
+ kill -9 $pid2
+ echo 'killed '${JAR_NAME}' process-->'$pid2
+ done
+fi
+done
diff --git a/internal-flume/docker-compose.yml b/internal-flume/origin/docker-compose.yml
index 8c1f330..8c1f330 100644
--- a/internal-flume/docker-compose.yml
+++ b/internal-flume/origin/docker-compose.yml
diff --git a/internal-flume/flumeController/bootstart.log b/internal-flume/origin/flumeController/bootstart.log
index 46952f0..46952f0 100644
--- a/internal-flume/flumeController/bootstart.log
+++ b/internal-flume/origin/flumeController/bootstart.log
diff --git a/internal-flume/flumeController/change_log_lv.sh b/internal-flume/origin/flumeController/change_log_lv.sh
index 77c9324..77c9324 100755
--- a/internal-flume/flumeController/change_log_lv.sh
+++ b/internal-flume/origin/flumeController/change_log_lv.sh
diff --git a/internal-flume/flumeController/clean_log_all.sh b/internal-flume/origin/flumeController/clean_log_all.sh
index 879e227..879e227 100755
--- a/internal-flume/flumeController/clean_log_all.sh
+++ b/internal-flume/origin/flumeController/clean_log_all.sh
diff --git a/internal-flume/flumeController/clean_start_all.sh b/internal-flume/origin/flumeController/clean_start_all.sh
index 2e0a735..2e0a735 100755
--- a/internal-flume/flumeController/clean_start_all.sh
+++ b/internal-flume/origin/flumeController/clean_start_all.sh
diff --git a/internal-flume/flumeController/count_all.sh b/internal-flume/origin/flumeController/count_all.sh
index c356af5..c356af5 100755
--- a/internal-flume/flumeController/count_all.sh
+++ b/internal-flume/origin/flumeController/count_all.sh
diff --git a/internal-flume/flumeController/start_all.sh b/internal-flume/origin/flumeController/start_all.sh
index ac4ae82..ac4ae82 100755
--- a/internal-flume/flumeController/start_all.sh
+++ b/internal-flume/origin/flumeController/start_all.sh
diff --git a/internal-flume/flumeController/stop_all.sh b/internal-flume/origin/flumeController/stop_all.sh
index 80320de..80320de 100755
--- a/internal-flume/flumeController/stop_all.sh
+++ b/internal-flume/origin/flumeController/stop_all.sh
diff --git a/internal-flume/flumeMonitor/file-monitor-flume.log b/internal-flume/origin/flumeMonitor/file-monitor-flume.log
index e69de29..e69de29 100644
--- a/internal-flume/flumeMonitor/file-monitor-flume.log
+++ b/internal-flume/origin/flumeMonitor/file-monitor-flume.log
diff --git a/internal-flume/flumeMonitor/flumeConfList b/internal-flume/origin/flumeMonitor/flumeConfList
index 29554b4..29554b4 100644
--- a/internal-flume/flumeMonitor/flumeConfList
+++ b/internal-flume/origin/flumeMonitor/flumeConfList
diff --git a/internal-flume/flumeMonitor/readMe.txt b/internal-flume/origin/flumeMonitor/readMe.txt
index 3f19da1..3f19da1 100644
--- a/internal-flume/flumeMonitor/readMe.txt
+++ b/internal-flume/origin/flumeMonitor/readMe.txt
diff --git a/internal-flume/flumeMonitor/taskFlumeMinSum.sh b/internal-flume/origin/flumeMonitor/taskFlumeMinSum.sh
index 36492f1..36492f1 100755
--- a/internal-flume/flumeMonitor/taskFlumeMinSum.sh
+++ b/internal-flume/origin/flumeMonitor/taskFlumeMinSum.sh