案例:将nginx的日志实时传输到hdfs和kafka上

案例:将nginx的日志实时传输到hdfs和kafka上

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
a1.sources = r1
a1.channels = c1 c2
a1.sinks = k1 k2

a1.sources.r1.type=taildir
a1.sources.r1.filegroups = f1
a1.sources.r1.filegroups.f1=/usr/local/nginx/logs/access.log

a1.channels.c1.type=memory
a1.channels.c1.capacity=10000
a1.channels.c1.transactionCapacity=100

a1.channels.c2.type=memory
a1.channels.c2.capacity=10000
a1.channels.c2.transactionCapacity=100

a1.sinks.k1.type = hdfs
a1.sinks.k1.hdfs.path=/data/hkjcpdd/%Y%m%d
a1.sinks.k1.hdfs.useLocalTimeStamp=true
a1.sinks.k1.hdfs.rollInterval=10
a1.sinks.k1.hdfs.fileType=DataStream

a1.sinks.k2.type = org.apache.flume.sink.kafka.KafkaSink
a1.sinks.k2.kafka.bootstrap.servers = master:9092
a1.sinks.k2.kafka.topic = hkjcpdd

a1.sources.r1.channels = c1
a1.sinks.k1.channel = c1
a1.sinks.k2.channel = c1