原文地址:
简单来讲他具体的工作流程就是 logstash agent 监控并过滤日志,将过滤后的日志内容发给redis(这里的redis只处理队列不做存储),logstash index将日志收集在一起交给全文搜索服务ElasticSearch 可以用ElasticSearch进行自定义搜索 通过Kibana 来结合 自定义搜索进行页面展示
- ruby 运行Kibana 必须
- rubygems 安装ruby扩展必须
- bundler 功能类似于yum
- JDK 运行java程序必须
- redis 用来处理日志队列
- logstash 收集、过滤日志
- ElasticSearch 全文搜索服务(logstash集成了一个)
kibana 页面展示
首先到 logstash index服务器上面,logstash分为 index和aget ,agent负责监控、过滤日志,index负责收集日志并将日志交给ElasticSearch 做搜索此外 logstash 的收集方式分为 standalone 和 centralized。
standalone 是所有功能都在一个服务器上面,自发自收,centralized 就是集中收集,一台服务器接收所有shipper(个人理解就是logstash agent)的日志。
其实 logstash本身不分 什么 shipper 和 collector ,只不过就是配置文件不同而已,我们这次按照集中的方式来测试
这里有两台服务器
192.168.124.128 logstash index,ElasticSearch,kibana,JDK
192.168.124.132 logstash agent,redis,JDK准备工作
安装:openssl
卸载旧版本
apt-get remove opensslapt-get autoremove openssl
下载最新版本
wget
tar -zxvf openssl-1.0.1i.tar.gzcd /opt/openssl-1.0.1i./config --prefix=/usr/local/sslmake & make install
建立软连接
ln -s /usr/local/ssl/bin/openssl /usr/bin/opensslln -s /usr/local/ssl/include/openssl /usr/include/openssl
刷新动态配置
vim /etc/ld.so.conf
在文末插入一行
/usr/local/ssl/libldconfig -v
测试
openssl version -a
安装PCRE库
wget
tar -zxvf pcre-8.33.tar.gzcd pcre-8.33./configure --prefix=/usr/local/pcre-8.33make & make install
安装zlib
wget
tar -zxvf zlib-1.2.8.tar.gzcd zlib-1.2.8./configure --prefix=/usr/local/zlib-1.2.8make & make install
安装nginx
wget
tar -zxvf nginx-1.6.1.tar.gzcd nginx-1.6.1./configure --prefix=/usr/local/nginx --with-http_ssl_module --with-openssl=/opt/openssl-1.0.1i --with-pcre=/opt/pcre-8.33 --with-zlib=/opt/zlib-1.2.8
nginx 命令
启动:/usr/local/nginx/sbin/nginx重启:/usr/local/nginx/sbin/nginx –s reload停止:/usr/local/nginx/sbin/nginx -s stop查看主进程:netstat -ntlp检查是否启动成功:netstat -ano|grep 80
安装ruby 运行Kibana 必须
sudo apt-get update wget http://cache.ruby-lang.org/pub/ruby/2.1/ruby-2.1.2.tar.gz./configure --prefix=/usr/local/rubymake && make install
环境设置
vi /etc/environment
将Ruby的路径加入环境变量 中并保存/etc/environment,如下面内容:
PATH="/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/ruby/bin"
修改了环境变量文件后,需要通过source命令让修改马上生效,命令如下:
$ source /etc/environment
为了检查安装是否成功,可以输入下面的命令进行测试 :
$ruby –v
确认安装成功后通过一下命令添加命令链接,目前我也不清楚创建这些链接的目的是什么,按照Ruby“约定大于配置”的原则,应该是一种约定。(keyboardota)
$ sudo ln -s /usr/local/ruby/bin/ruby /usr/local/bin/ruby$ sudo ln -s /usr/local/ruby/bin/gem /usr/bin/gem
或者:
apt-get install ruby-full
安装rubygems ruby扩展必须
wget
tar -zxvf rubygems-2.4.1.tgzcd rubygems-2.4.1ruby setup.rb
安装redis 用来处理日志队列
wget
tar -zxvf redis-2.8.13.tar.gzcd redis-2.8.13makevim redis.conf设置 "daemonize yes"启动:/usr/local/redis-2.8.13/src/redis-server /usr/local/redis-2.8.13/redis.conf
安装 elasticsearch 全文搜索服务(logstash集成了一个)
wget
tar -zxvf elasticsearch-1.3.2.tar.gzcd elasticsearch-1.3.2启动: /usr/local/elasticsearch-1.3.2/bin/elasticsearch -d访问 http://localhost:9200
安装:logstash 收集、过滤日志
wget
tar -zxvf logstash-1.4.2.tar.gz
启动
nohup /usr/local/logstash-1.4.2/bin/logstash -f /usr/local/logstash-1.4.2/agent.conf &
nohup /usr/local/logstash-1.4.2/bin/logstash -f /usr/local/logstash-1.4.2/indexer.conf &
vim /usr/local/logstash-1.4.2/agent.conf
input { file { path => [ "/var/log/*.log", "/var/log/messages", "/var/log/syslog", "/var/log/denyhosts", "/var/log/dmesg", "/var/log/faillog", "/var/log/aptitude" ] start_position => beginning } file { type => "nginx-access" path => "/var/log/nginx/access.log" }}output { redis{ host =>"192.168.124.128" data_type => "list" key => "logstash" }}
vim /usr/local/logstash-1.4.2/indexer.conf
input { redis { host => "192.168.124.128" data_type => "list" key => "logstash" }}output { elasticsearch { host => "192.168.124.132" #指定elasticsearch服务位置 }}
安装Kibana
wget
tar -zxvf kibana-3.1.0.tar.gzvim /usr/local/kibana-3.1.0/config/ 可以通过whereis kibana找到kibana具体物理地址 在config目录下有个kibana.yml配置文件
搜索"elasticsearch"参数,并对其进行修改以适应您的环境:
elasticsearch: "http://192.168.124.132:9200",
您还可以修改default_route参数,默认打开logstash仪表板而不是Kibana欢迎页面:
default_route : '/dashboard/file/logstash.json', (这个在哪里配置,没看到...)
下载配置模板
wget
修改Nginx配置
vim /usr/local/nginx/conf/nginx.conf
增加Server节点
# # Nginx proxy for Elasticsearch + Kibana # # In this setup, we are password protecting the saving of dashboards. You may # wish to extend the password protection to all paths. # # Even though these paths are being called as the result of an ajax request, the # browser will prompt for a username/password on the first request # # If you use this, you'll want to point config.js at http://FQDN:80/ instead of # http://FQDN:9200 # server { listen *:80 ; server_name localhost; access_log /usr/local/nginx/logs/kibana.access.log; location / { root /usr/local/kibana-3.1.0; index index.html index.htm; } location ~ ^/_aliases$ { proxy_pass http://127.0.0.1:9200; proxy_read_timeout 90; } location ~ ^/.*/_aliases$ { proxy_pass http://127.0.0.1:9200; proxy_read_timeout 90; } location ~ ^/_nodes$ { proxy_pass http://127.0.0.1:9200; proxy_read_timeout 90; } location ~ ^/.*/_search$ { proxy_pass http://127.0.0.1:9200; proxy_read_timeout 90; } location ~ ^/.*/_mapping { proxy_pass http://127.0.0.1:9200; proxy_read_timeout 90; } # Password protected end points location ~ ^/kibana-int/dashboard/.*$ { proxy_pass http://127.0.0.1:9200; proxy_read_timeout 90; limit_except GET { proxy_pass http://127.0.0.1:9200; auth_basic "Restricted"; auth_basic_user_file /usr/local/nginx/kibana.myhost.org.htpasswd; } } location ~ ^/kibana-int/temp.*$ { proxy_pass http://127.0.0.1:9200; proxy_read_timeout 90; limit_except GET { proxy_pass http://127.0.0.1:9200; auth_basic "Restricted"; auth_basic_user_file /usr/local/nginx/kibana.myhost.org.htpasswd; } } }
如果有防火墙需要放开这些端口:
- port 80 (for the web interface)
- port 5544 (to receive remote syslog messages)
- port 6379 (for the redis broker)
- port 9200 (so the web interface can access elasticsearch)