在看大型网站的中间件技术,对于Elasticsearch有点趣味,所以将配置流程记录了一下。
为什么要用ELK
ELK实际上是三个工具,Elastricsearch + LogStash + Kibana,通过ELK,用来收集日志还有进行日志剖析,最初通过可视化UI进行展现。一开始业务量比拟小的时候,通过简略的SLF4J+Logger在服务器打印日志,通过grep进行简略查问,然而随着业务量减少,数据量也会一直减少,所以应用ELK能够进行大数量的日志收集和剖析
简略画了一下架构图
在环境配置中,次要介绍Mac和linux配置,windows零碎大致相同,当然,前提是大家都装置了JDK1.8及以上版本~
[root@VM_234_23_centos ~]# java -versionjava version "1.8.0_161"Java(TM) SE Runtime Environment (build 1.8.0_161-b12)Java HotSpot(TM) 64-Bit Server VM (build 25.161-b12, mixed mode)
留神:
高版本的ELK同样须要高版本的JDK反对,本文配置的ELK版本是6.0+,所以须要的JDK版本不小于1.8
ElasticSearch
Elasticsearch 是一个分布式的 RESTful 格调的搜寻和数据分析引擎,可能解决不断涌现出的各种用例。作为 Elastic Stack 的外围,它集中存储您的数据,帮忙您发现意料之中以及意料之外的状况。
Mac装置和运行
装置:brew install elasticsearch运行:elasticsearch
linux: 从Elasticsearch官网地址下载(也能够下载完,通过ftp之类的工具传上去),gz文件的话通过tar进行解压缩,而后进入bin目录下运行软件
[root@VM_234_23_centos app]# curl -L -O https://artifacts.elastic.co/downloads/elasticsearch/elasticsearch-6.2.4.tar.gz[root@VM_234_23_centos app]# tar -zxvf elasticsearch-6.2.4.tar.gz[root@VM_234_23_centos app]# cd elasticsearch-6.2.4[root@VM_234_23_centos elasticsearch-6.2.4]# ./bin/elasticsearch
留神:
在Linux机器上,运行elasticsearch须要一个新的用户组,文章最初有Elastic在linux装置的踩坑记录
Logstash
Logstash 是开源的服务器端数据处理管道,可能同时从多个起源采集数据,转换数据,而后将数据发送到您最喜爱的 “存储库” 中。(咱们的存储库当然是 Elasticsearch。)-官网卖萌
1.软件装置
Mac装置:
brew install logstash
linux装置:
[root@VM_234_23_centos app]# curl -L -O https://artifacts.elastic.co/downloads/logstash/logstash-6.3.2.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed100 137M 100 137M 0 0 5849k 0 0:00:24 0:00:24 --:--:-- 6597k[root@VM_234_23_centos app]# tar -zxvf logstash-6.3.2.tar.gz
2.批改配置文件
vim /etc/logstash.conf
conf文件,指定要应用的插件,和配置对应的elasticsearch的hosts
input { stdin { } }output { elasticsearch { hosts => ["localhost:9200"] } stdout { codec => rubydebug }}
3.运行
bin/logstash -f logstash.conf
4.拜访http://localhost:9600/
{ "host": "=-=", "version": "6.2.4", "http_address": "127.0.0.1:9600", "id": "5b47e81f-bdf8-48fc-9537-400107a13bd2", "name": "=-=", "build_date": "2018-04-12T22:29:17Z", "build_sha": "a425a422e03087ac34ad6949f7c95ec6d27faf14", "build_snapshot": false}
在elasticsearch日志中,也能看到logstash失常退出的日志
[2018-08-16T14:08:36,436][INFO ][o.e.c.m.MetaDataIndexTemplateService] [f2s1SD8] adding template [logstash] for index patterns [logstash-*]
看到这种返回值,示意曾经胜利装置和启动
踩坑:
_在运行的那一步,有可能遇到内存调配谬误
Java HotSpot™ 64-Bit Server VM warning: INFO: os::commit_memory(0x00000000c5330000, 986513408, 0) failed; error=’Cannot allocate memory’ (errno=12)
这个谬误很显著就是内存不足,因为集体购买的是腾讯云1G内存的服务器(如果是壕,请随便购买更高的配置=-=),曾经运行了elasticsearch,导致logstash调配不到足够的内存,所以最初要批改一下jvm配置。_
[root@VM_234_23_centos logstash-6.3.2]# cd config/[root@VM_234_23_centos config]# lltotal 28-rw-r--r-- 1 root root 1846 Jul 20 14:19 jvm.options-rw-r--r-- 1 root root 4466 Jul 20 14:19 log4j2.properties-rw-r--r-- 1 root root 8097 Jul 20 14:19 logstash.yml-rw-r--r-- 1 root root 3244 Jul 20 14:19 pipelines.yml-rw-r--r-- 1 root root 1696 Jul 20 14:19 startup.options[root@VM_234_23_centos config]# vim jvm.options
将-Xms1g -Xmx1g批改为
-Xms256m -Xmx256m
而后就能失常启动了~~
Kibana
1.软件装置
Kibana 让您可能可视化 Elasticsearch 中的数据并操作 Elastic Stack,因而您能够在这里解开任何疑难:例如,为何会在凌晨 2:00 被传呼,雨水会对季度数据造成怎么的影响。(而且展现的图标非常酷炫)
Mac装置
brew install kibana
linux装置,官网下载地址
[root@VM_234_23_centos app]# curl -L -O https://artifacts.elastic.co/downloads/kibana/kibana-6.3.2-linux-x86_64.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 195M 0 271k 0 0 19235 0 2:57:54 0:00:14 2:57:40 26393
在这一步,有可能下载速度奇慢,所以我本地下载好之后,通过rz命令传输到服务器
[root@VM_234_23_centos app]# rzrz waiting to receive.Starting zmodem transfer. Press Ctrl+C to cancel.Transferring kibana-6.3.2-linux-x86_64.tar.gz... 100% 200519 KB 751 KB/sec 00:04:27 0 Errors [root@VM_234_23_centos app]# tar -zxvf kibana-6.3.2-linux-x86_64.tar.gz
2.批改配置
批改 config/kibana.yml 配置文件,设置 elasticsearch.url 指向 Elasticsearch 实例。
如果跟我一样应用默认的配置,能够不须要批改该文件
3.启动
[root@VM_234_23_centos kibana]# ./bin/kibana
4.拜访 http://localhost:5601/app/kib...
界面显示了这么多功能,上面通过整合SLF4J+LogBack
整合Spring+Logstash
1.批改logstash.conf后,重新启动logstash
input { # stdin { } tcp { # host:port就是下面appender中的 destination, # 这里其实把logstash作为服务,开启9250端口接管logback收回的音讯 host => "127.0.0.1" port => 9250 mode => "server" tags => ["tags"] codec => json_lines }}output { elasticsearch { hosts => ["localhost:9200"] } stdout { codec => rubydebug }}
2.在Java利用中援用依赖
<dependency> <groupId>net.logstash.logback</groupId> <artifactId>logstash-logback-encoder</artifactId> <version>5.2</version></dependency>
3.在Logback.xml中配置日志输入
<!--日志导出的到 Logstash--><appender name="stash" class="net.logstash.logback.appender.LogstashTcpSocketAppender"> <destination>localhost:9250</destination> <!-- encoder必须配置,有多种可选 --> <encoder charset="UTF-8" class="net.logstash.logback.encoder.LogstashEncoder" > <!-- "appname":"ye_test" 的作用是指定创立索引的名字时用,并且在生成的文档中会多了这个字段 --> <customFields>{"appname":"ye_test"}</customFields> </encoder></appender> <root level="INFO"> <appender-ref ref="stash"/></root>
因为我在第一步骤中,没有指定对应的index,所以在服务启动的时候,日志采集器Logstash帮我主动创立了logstash-timestamp的index。
4.在kibana中增加index索引
5.在右边discover中查看索引信息
6.增加可视化图表Visualize
还有更多功能还在摸索中,首先环境搭起来才会用能源持续去学习~
踩坑记录
启动报错
uncaught exception in thread [main] org.elasticsearch.bootstrap.StartupException: java.lang.RuntimeException: can not run elasticsearch as root
起因:不能应用Root权限登录
解决方案:切换用户
[root@VM_234_23_centos ~]# groupadd es[root@VM_234_23_centos ~]# useradd es -g es -p es[root@VM_234_23_centos ~]# chown es:es /home/app/elasticsearch/# 切换用户,记得su - ,这样能力取得环境变量[root@VM_234_23_centos ~]# sudo su - es
Exception in thread “main” java.nio.file.AccessDeniedException:
谬误起因:应用非 root用户启动ES,而该用户的文件权限有余而被拒绝执行。
解决办法:chown -R 用户名:用户名 文件(目录)名
例如:chown -R abc:abc searchengine 再启动ES就失常了
elasticsearch启动后报Killed
[2018-07-13T10:19:44,775][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [aggs-matrix-stats][2018-07-13T10:19:44,779][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [analysis-common][2018-07-13T10:19:44,780][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [ingest-common][2018-07-13T10:19:44,780][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [lang-expression][2018-07-13T10:19:44,780][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [lang-mustache][2018-07-13T10:19:44,780][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [lang-painless][2018-07-13T10:19:44,780][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [mapper-extras][2018-07-13T10:19:44,780][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [parent-join][2018-07-13T10:19:44,780][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [percolator][2018-07-13T10:19:44,780][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [rank-eval][2018-07-13T10:19:44,781][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [reindex][2018-07-13T10:19:44,781][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [repository-url][2018-07-13T10:19:44,781][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [transport-netty4][2018-07-13T10:19:44,781][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [tribe][2018-07-13T10:19:44,781][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [x-pack-core][2018-07-13T10:19:44,781][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [x-pack-deprecation][2018-07-13T10:19:44,781][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [x-pack-graph][2018-07-13T10:19:44,781][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [x-pack-logstash][2018-07-13T10:19:44,782][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [x-pack-ml][2018-07-13T10:19:44,782][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [x-pack-monitoring][2018-07-13T10:19:44,782][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [x-pack-rollup][2018-07-13T10:19:44,782][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [x-pack-security][2018-07-13T10:19:44,782][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [x-pack-sql][2018-07-13T10:19:44,782][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [x-pack-upgrade][2018-07-13T10:19:44,782][INFO ][o.e.p.PluginsService ] [f2s1SD8] loaded module [x-pack-watcher][2018-07-13T10:19:44,783][INFO ][o.e.p.PluginsService ] [f2s1SD8] no plugins loadedKilled
批改config目录下的jvm.options,将堆的大小设置小一点
# Xms represents the initial size of total heap space# Xmx represents the maximum size of total heap space-Xms512m-Xmx512m
虚拟内存有余
max virtual memory areas vm.max_map_count [65530] is too low, increase to at least [262144]
[2018-07-13T14:02:06,749][DEBUG][o.e.a.ActionModule ] Using REST wrapper from plugin org.elasticsearch.xpack.security.Security[2018-07-13T14:02:07,249][INFO ][o.e.d.DiscoveryModule ] [f2s1SD8] using discovery type [zen][2018-07-13T14:02:09,173][INFO ][o.e.n.Node ] [f2s1SD8] initialized[2018-07-13T14:02:09,174][INFO ][o.e.n.Node ] [f2s1SD8] starting ...[2018-07-13T14:02:09,539][INFO ][o.e.t.TransportService ] [f2s1SD8] publish_address {10.105.234.23:9300}, bound_addresses {0.0.0.0:9300}[2018-07-13T14:02:09,575][INFO ][o.e.b.BootstrapChecks ] [f2s1SD8] bound or publishing to a non-loopback address, enforcing bootstrap checksERROR: [1] bootstrap checks failed[1]: max virtual memory areas vm.max_map_count [65530] is too low, increase to at least [262144][2018-07-13T14:02:09,621][INFO ][o.e.n.Node ] [f2s1SD8] stopping ...[2018-07-13T14:02:09,726][INFO ][o.e.n.Node ] [f2s1SD8] stopped[2018-07-13T14:02:09,726][INFO ][o.e.n.Node ] [f2s1SD8] closing ...[2018-07-13T14:02:09,744][INFO ][o.e.n.Node ] [f2s1SD8] closed
须要批改虚拟内存的大小(在root权限下)
[root@VM_234_23_centos elasticsearch]# vim /etc/sysctl.conf# 插入下列代码后保留退出vm.max_map_count=655360[root@VM_234_23_centos elasticsearch]# sysctl -p# 最初重启elastricsearch
作者:JingQ
起源:https://www.sevenyuan.cn