# Kubernetes Cluster installation
This article mainly introduces how to use helm or kubectl to deploy ClickVisual to kubernetes cluster.
# 1. Deployment requirements
- Kubernetes >= 1.17
- Helm >= 3.0.0
# 2. Deploy fluent bit (Reference)
You can directly refer to the official fluent bit website for deployment https://docs.fluentbit.io/ , just ensure that the data written to Kafka contains the following two fields.
- time
- log
https://docs.fluentbit.io/manual/installation/kubernetes#installation (opens new window)
Fluent Bit Kubernetes Daemonset https://github.com/fluent/fluent-bit-kubernetes-logging (opens new window)
fluentbit-configmap.yaml is as follows:
apiVersion: v1
kind: ConfigMap
metadata:
name: fluent-bit-config
namespace: kube-system
labels:
k8s-app: fluent-bit
data:
# Configuration files: server, input, filters and output
# ======================================================
fluent-bit.conf: |
[SERVICE]
Flush 1
Log_Level info
Daemon off
Parsers_File parsers.conf
HTTP_Server On
HTTP_Listen 0.0.0.0
HTTP_Port 2020
@INCLUDE input-kubernetes.conf
@INCLUDE filter-kubernetes.conf
@INCLUDE filter-modify.conf
@INCLUDE output-kafka.conf
# Disabled when env is configured in DaemonSet
#@Set CLUSTER_NAME=shimodev
#@Set KAFKA_BROKERS=127.0.0.1:9092
input-kubernetes.conf: |
[INPUT]
Name tail
# Tag identifies the data source, which is used to select data in the subsequent process Filter and Output
Tag ingress.*
Path /var/log/containers/nginx-ingress-controller*.log
Parser docker
DB /var/log/flb_ingress.db
Mem_Buf_Limit 15MB
Buffer_Chunk_Size 32k
Buffer_Max_Size 64k
# Skip the row when length greater than Buffer_Max_Size. If Skip_Long_Lines is Off, the collection will be stopped when encounter the row above.
Skip_Long_Lines On
Refresh_Interval 10
# If the acquisition file has no database offset record, it is read from the header of the file. When the log file is large, it will lead to the increase of fluent memory usage and oomkill
#Read_from_Head On
[INPUT]
Name tail
# Tag identifies the data source, which is used to select data in the subsequent process Filter and Output
Tag ingress_stderr.*
Path /var/log/containers/nginx-ingress-controller*.log
Parser docker
DB /var/log/flb_ingress_stderr.db
Mem_Buf_Limit 15MB
Buffer_Chunk_Size 32k
Buffer_Max_Size 64k
# Skip the row when length greater than Buffer_Max_Size. If Skip_Long_Lines is Off, the collection will be stopped when encounter the row above.
Skip_Long_Lines On
Refresh_Interval 10
# If the acquisition file has no database offset record, it is read from the header of the file. When the log file is large, it will lead to the increase of fluent memory usage and oomkill
#Read_from_Head On
[INPUT]
Name tail
Tag kube.*
Path /var/log/containers/*_default_*.log,/var/log/containers/*_release_*.log
Exclude_path *fluent-bit-*,*mongo-*,*minio-*,*mysql-*
Parser docker
DB /var/log/flb_kube.db
Mem_Buf_Limit 15MB
Buffer_Chunk_Size 1MB
Buffer_Max_Size 5MB
# Skip the row when length greater than Buffer_Max_Size. If Skip_Long_Lines is Off, the collection will be stopped when encounter the row above.
Skip_Long_Lines On
Refresh_Interval 10
[INPUT]
Name tail
Tag ego.*
Path /var/log/containers/*_default_*.log,/var/log/containers/*_release_*.log
Exclude_path *fluent-bit-*,*mongo-*,*minio-*,*mysql-*
Parser docker
DB /var/log/flb_ego.db
Mem_Buf_Limit 15MB
Buffer_Chunk_Size 1MB
Buffer_Max_Size 5MB
Skip_Long_Lines On
Refresh_Interval 10
filter-kubernetes.conf: |
[FILTER]
Name kubernetes
Match ingress.*
Kube_URL https://kubernetes.default.svc:443
Kube_CA_File /var/run/secrets/kubernetes.io/serviceaccount/ca.crt
Kube_Token_File /var/run/secrets/kubernetes.io/serviceaccount/token
Kube_Tag_Prefix ingress.var.log.containers.
# When enabled,it parses the JSON content of the log field, extract it to the root level, and attach it to the field specified by Merge_Log_Key.
Merge_Log Off
#Merge_Log_Key log_processed
#Merge_Log_Trim On
# Whether to keep the original log field after merging the log field
Keep_Log On
K8S-Logging.Parser On
K8S-Logging.Exclude Off
Labels Off
Annotations Off
#Regex_Parser
[FILTER]
Name kubernetes
Match ingress_stderr.*
Kube_URL https://kubernetes.default.svc:443
Kube_CA_File /var/run/secrets/kubernetes.io/serviceaccount/ca.crt
Kube_Token_File /var/run/secrets/kubernetes.io/serviceaccount/token
Kube_Tag_Prefix ingress_stderr.var.log.containers.
# Merge_Log=On Parse the JSON content of the log field, extract it to the root level, and attach it to the field specified by Merge_Log_Key.
Merge_Log Off
# Whether to keep the original log field after merging the log field
Keep_Log Off
K8S-Logging.Parser On
K8S-Logging.Exclude Off
Labels Off
Annotations Off
#Regex_Parser
[FILTER]
Name kubernetes
Match kube.*
Kube_URL https://kubernetes.default.svc:443
Kube_CA_File /var/run/secrets/kubernetes.io/serviceaccount/ca.crt
Kube_Token_File /var/run/secrets/kubernetes.io/serviceaccount/token
Kube_Tag_Prefix kube.var.log.containers.
Merge_Log Off
Keep_Log On
K8S-Logging.Parser On
K8S-Logging.Exclude Off
Labels Off
Annotations Off
[FILTER]
Name kubernetes
Match ego.*
Kube_URL https://kubernetes.default.svc:443
Kube_CA_File /var/run/secrets/kubernetes.io/serviceaccount/ca.crt
Kube_Token_File /var/run/secrets/kubernetes.io/serviceaccount/token
Kube_Tag_Prefix ego.var.log.containers.
Merge_Log Off
Keep_Log Off
K8S-Logging.Parser On
K8S-Logging.Exclude Off
Labels Off
Annotations Off
filter-modify.conf: |
[FILTER]
Name nest
Match *
Wildcard pod_name
Operation lift
Nested_under kubernetes
Add_prefix kubernetes_
[FILTER]
Name modify
Match *
Rename time _time_
Rename log _log_
Rename stream _source_
Rename kubernetes_host _node_name_
Rename kubernetes_namespace_name _namespace_
Rename kubernetes_container_name _container_name_
Rename kubernetes_pod_name _pod_name_
Remove kubernetes_pod_id
Remove kubernetes_docker_id
Remove kubernetes_container_hash
Remove kubernetes_container_image
Add _cluster_ ${CLUSTER_NAME}
Add _log_agent_ ${HOSTNAME}
# ${NODE_IP} Configure ENV injection through the daemonset
Add _node_ip_ ${NODE_IP}
[FILTER]
Name grep
Match ingress.*
#Regex container_name ^nginx-ingress-controller$
#Regex stream ^stdout$
Exclude _source_ ^stderr$
# Exclude TCP agent logs (different log formats affect Collection)
Exclude log ^\[*
[FILTER]
Name grep
Match ingress_stderr.*
Exclude _source_ ^stdout$
[FILTER]
Name grep
Match kube.*
#Regex stream ^stdout$
Exclude log (ego.sys)
[FILTER]
Name grep
Match ego.*
#Regex lname ^(ego.sys)$
Regex log ("lname":"ego.sys")
# [FILTER]
# Name modify
# Match ego.*
# Hard_rename ts _time_
output-kafka.conf: |
[OUTPUT]
Name kafka
Match ingress.*
Brokers ${KAFKA_BROKERS}
Topics ingress-stdout-logs-${CLUSTER_NAME}
#Timestamp_Key @timestamp
Timestamp_Key _time_
Retry_Limit false
# hides errors "Receive failed: Disconnected" when kafka kills idle connections
rdkafka.log.connection.close false
# producer buffer is not included in http://fluentbit.io/documentation/0.12/configuration/memory_usage.html#estimating
rdkafka.queue.buffering.max.kbytes 10240
# for logs you'll probably want this ot be 0 or 1, not more
rdkafka.request.required.acks 1
[OUTPUT]
Name kafka
Match ingress_stderr.*
Brokers ${KAFKA_BROKERS}
Topics ingress-stderr-logs-${CLUSTER_NAME}
#Timestamp_Key @timestamp
Timestamp_Key _time_
Retry_Limit false
# hides errors "Receive failed: Disconnected" when kafka kills idle connections
rdkafka.log.connection.close false
# producer buffer is not included in http://fluentbit.io/documentation/0.12/configuration/memory_usage.html#estimating
rdkafka.queue.buffering.max.kbytes 10240
# for logs you'll probably want this ot be 0 or 1, not more
rdkafka.request.required.acks 1
[OUTPUT]
Name kafka
Match kube.*
Brokers ${KAFKA_BROKERS}
Topics app-stdout-logs-${CLUSTER_NAME}
Timestamp_Key _time_
Retry_Limit false
# hides errors "Receive failed: Disconnected" when kafka kills idle connections
rdkafka.log.connection.close false
# producer buffer is not included in http://fluentbit.io/documentation/0.12/configuration/memory_usage.html#estimating
rdkafka.queue.buffering.max.kbytes 10240
# for logs you'll probably want this ot be 0 or 1, not more
rdkafka.request.required.acks 1
[OUTPUT]
Name kafka
Match ego.*
Brokers ${KAFKA_BROKERS}
Topics ego-stdout-logs-${CLUSTER_NAME}
Timestamp_Key _time_
Retry_Limit false
# hides errors "Receive failed: Disconnected" when kafka kills idle connections
rdkafka.log.connection.close false
# producer buffer is not included in http://fluentbit.io/documentation/0.12/configuration/memory_usage.html#estimating
rdkafka.queue.buffering.max.kbytes 10240
# for logs you'll probably want this ot be 0 or 1, not more
rdkafka.request.required.acks 1
parsers.conf: |
[PARSER]
Name apache
Format regex
Regex ^(?<host>[^ ]*) [^ ]* (?<user>[^ ]*) \[(?<time>[^\]]*)\] "(?<method>\S+)(?: +(?<path>[^\"]*?)(?: +\S*)?)?" (?<code>[^ ]*) (?<size>[^ ]*)(?: "(?<referer>[^\"]*)" "(?<agent>[^\"]*)")?$
Time_Key time
Time_Format %d/%b/%Y:%H:%M:%S %z
[PARSER]
Name apache2
Format regex
Regex ^(?<host>[^ ]*) [^ ]* (?<user>[^ ]*) \[(?<time>[^\]]*)\] "(?<method>\S+)(?: +(?<path>[^ ]*) +\S*)?" (?<code>[^ ]*) (?<size>[^ ]*)(?: "(?<referer>[^\"]*)" "(?<agent>[^\"]*)")?$
Time_Key time
Time_Format %d/%b/%Y:%H:%M:%S %z
[PARSER]
Name apache_error
Format regex
Regex ^\[[^ ]* (?<time>[^\]]*)\] \[(?<level>[^\]]*)\](?: \[pid (?<pid>[^\]]*)\])?( \[client (?<client>[^\]]*)\])? (?<message>.*)$
[PARSER]
Name nginx
Format regex
Regex ^(?<remote>[^ ]*) (?<host>[^ ]*) (?<user>[^ ]*) \[(?<time>[^\]]*)\] "(?<method>\S+)(?: +(?<path>[^\"]*?)(?: +\S*)?)?" (?<code>[^ ]*) (?<size>[^ ]*)(?: "(?<referer>[^\"]*)" "(?<agent>[^\"]*)")?$
Time_Key time
Time_Format %d/%b/%Y:%H:%M:%S %z
[PARSER]
Name json
Format json
Time_Key time
Time_Format %d/%b/%Y:%H:%M:%S %z
[PARSER]
Name docker
Format json
Time_Key time
Time_Format %Y-%m-%dT%H:%M:%S.%L
Time_Keep On
# Similar to the effect Merge_Log=On at filter stage, the JSON content of the log field is parsed, but cannot be extracted to the root level
#Decode_Field_As escaped_utf8 kubernetes do_next
#Decode_Field_As json kubernetes
[PARSER]
# http://rubular.com/r/tjUt3Awgg4
Name cri
Format regex
Regex ^(?<time>[^ ]+) (?<stream>stdout|stderr) (?<logtag>[^ ]*) (?<message>.*)$
Time_Key time
Time_Format %Y-%m-%dT%H:%M:%S.%L%z
[PARSER]
Name syslog
Format regex
Regex ^\<(?<pri>[0-9]+)\>(?<time>[^ ]* {1,2}[^ ]* [^ ]*) (?<host>[^ ]*) (?<ident>[a-zA-Z0-9_\/\.\-]*)(?:\[(?<pid>[0-9]+)\])?(?:[^\:]*\:)? *(?<message>.*)$
Time_Key time
Time_Format %b %d %H:%M:%S
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
# Deploy ClickVisual
Clone git repository:
git clone https://github.com/clickvisual/clickvisual.git
# Use custom configuration
cd clickvisual && cp config/default.toml data/helm/clickvisual/default.toml
Edit mysql、auth and other segment configurations in data/helm/clickvisual/default.toml,update mysql.dsn,auth.redisAddr,auth.redisPassword as you want.
Update the value configs/default.toml
in data/helm/clickvisual/templates/deployment.yaml if you need
- name: EGO_CONFIG_PATH
value: "configs/default.toml"
2
# Install
Method 1: [recommended] install directly with Helm:
helm install clickvisual data/helm/clickvisual --set image.tag=latest --namespace default
If you have pushed the clickvisual image to your own harbor repository,use --set image.respository to change respository address.
helm install clickvisual data/helm/clickvisual --set image.repository=${YOUR_HARBOR}/${PATH}/clickvisual --set image.tag=latest --namespace default
Method 2: [optional] after rendering yaml with helm, install it manually through kubectl:
# Use helm template render clickvisual.yaml for install
helm template clickvisual data/helm/clickvisual --set image.tag=latest > clickvisual.yaml
# You can use "--set image.repository" to override the default image path
# helm template clickvisual clickvisual --set image.repository=${YOUR_HARBOR}/${PATH}/clickvisual --set image.tag=latest > clickvisual.yaml
# Check clickvisual.yaml and use kuebctl apply
kubectl apply -f clickvisual.yaml --namespace default
2
3
4
5
6
7
8