Open michal25 opened 4 years ago
Can you open the Kibana page? It seems it is unavailable ?
Kibana page results in white screen. No pictures, no text, no error message.
Ah, error message
Kibana server is not ready yet
Did you make sure the ngingx config is up to date - https://github.com/StamusNetworks/SELKS/wiki/Kibana-did-not-load-properly ?
Yes
I have exact this configuration file for nginx.
MV
Dne 26. 08. 20 v 21:19 Peter Manev napsal(a):
Did you make sure the ngingx config is up to date - https://github.com/StamusNetworks/SELKS/wiki/Kibana-did-not-load-properly ?
— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/StamusNetworks/SELKS/issues/251#issuecomment-681074319, or unsubscribe https://github.com/notifications/unsubscribe-auth/ABGUNEKOQI3VWOIVQSNPBMDSCVN4JANCNFSM4QLWP6DA.
-- Ing. Michal Vymazal Senior Cyber Security Architect
Linux Services CEO
vymazal@linuxservices.cz www.linuxservices.cz Office Computer
LinkedIn profile https://www.linkedin.com/in/linuxservices/
This mail can't contain any virus. I'm using only Open Source software.
Including the last two entrees ?
-- Regards, Peter Manev
On 27 Aug 2020, at 09:21, michal25 notifications@github.com wrote:
Yes
I have exact this configuration file for nginx.
MV
Dne 26. 08. 20 v 21:19 Peter Manev napsal(a):
Did you make sure the ngingx config is up to date - https://github.com/StamusNetworks/SELKS/wiki/Kibana-did-not-load-properly ?
— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/StamusNetworks/SELKS/issues/251#issuecomment-681074319, or unsubscribe https://github.com/notifications/unsubscribe-auth/ABGUNEKOQI3VWOIVQSNPBMDSCVN4JANCNFSM4QLWP6DA.
-- Ing. Michal Vymazal Senior Cyber Security Architect
Linux Services CEO
vymazal@linuxservices.cz www.linuxservices.cz Office Computer
LinkedIn profile https://www.linkedin.com/in/linuxservices/
This mail can't contain any virus. I'm using only Open Source software. — You are receiving this because you commented. Reply to this email directly, view it on GitHub, or unsubscribe.
Yes, exact copy of the configuration file.
M.
Dne 27. 08. 20 v 10:44 Peter Manev napsal(a):
Including the last two entrees ?
-- Regards, Peter Manev
On 27 Aug 2020, at 09:21, michal25 notifications@github.com wrote:
Yes
I have exact this configuration file for nginx.
MV
Dne 26. 08. 20 v 21:19 Peter Manev napsal(a):
Did you make sure the ngingx config is up to date -
https://github.com/StamusNetworks/SELKS/wiki/Kibana-did-not-load-properly ?
— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub
https://github.com/StamusNetworks/SELKS/issues/251#issuecomment-681074319, or unsubscribe
https://github.com/notifications/unsubscribe-auth/ABGUNEKOQI3VWOIVQSNPBMDSCVN4JANCNFSM4QLWP6DA.
-- Ing. Michal Vymazal Senior Cyber Security Architect
Linux Services CEO
vymazal@linuxservices.cz www.linuxservices.cz Office Computer
LinkedIn profile https://www.linkedin.com/in/linuxservices/
This mail can't contain any virus. I'm using only Open Source software. — You are receiving this because you commented. Reply to this email directly, view it on GitHub, or unsubscribe.
— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/StamusNetworks/SELKS/issues/251#issuecomment-681802658, or unsubscribe https://github.com/notifications/unsubscribe-auth/ABGUNEIG76YXMJ37RTRNTTLSCYMG5ANCNFSM4QLWP6DA.
-- Ing. Michal Vymazal Senior Cyber Security Architect
Linux Services CEO
vymazal@linuxservices.cz www.linuxservices.cz Office Computer
LinkedIn profile https://www.linkedin.com/in/linuxservices/
This mail can't contain any virus. I'm using only Open Source software.
I tryied to reset the SN dashboards, but this step terminates with error.
The kibana server is not available.
nginx configuration file is this https://github.com/StamusNetworks/SELKS/wiki/Kibana-did-not-load-properly
What is the output of
selks-health-check_stamus
?
root@SELKS2:~# selks-health-check_stamus
● suricata.service - LSB: Next Generation IDS/IPS
Loaded: loaded (/etc/init.d/suricata; generated)
Active: active (running) since Wed 2020-09-16 11:32:17 CEST; 59min ago
Docs: man:systemd-sysv-generator(8)
Tasks: 10 (limit: 4915)
Memory: 390.7M
CGroup: /system.slice/suricata.service
└─23105 /usr/bin/suricata -c /etc/suricata/suricata.yaml --pidfile /var/run/suricata.pid --af-packet -D -v --user=logstash
Sep 16 11:32:17 SELKS2 systemd[1]: Starting LSB: Next Generation IDS/IPS... Sep 16 11:32:17 SELKS2 suricata[23092]: Starting suricata in IDS (af-packet) mode... done. Sep 16 11:32:17 SELKS2 systemd[1]: Started LSB: Next Generation IDS/IPS. ● elasticsearch.service - Elasticsearch Loaded: loaded (/lib/systemd/system/elasticsearch.service; enabled; vendor preset: enabled) Active: active (running) since Wed 2020-09-16 11:14:15 CEST; 1h 17min ago Docs: https://www.elastic.co Main PID: 21914 (java) Tasks: 90 (limit: 4915) Memory: 4.7G CGroup: /system.slice/elasticsearch.service ├─21914 /usr/share/elasticsearch/jdk/bin/java -Xshare:auto -Des.networkaddress.cache.ttl=60 -Des.networkaddress.cache.negative.ttl=10 … └─22098 /usr/share/elasticsearch/modules/x-pack-ml/platform/linux-x86_64/bin/controller
Sep 16 11:14:05 SELKS2 systemd[1]: Starting Elasticsearch... Sep 16 11:14:15 SELKS2 systemd[1]: Started Elasticsearch. ● logstash.service - logstash Loaded: loaded (/etc/systemd/system/logstash.service; enabled; vendor preset: enabled) Active: active (running) since Wed 2020-08-26 12:45:26 CEST; 2 weeks 6 days ago Main PID: 24797 (java) Tasks: 42 (limit: 4915) Memory: 763.9M CGroup: /system.slice/logstash.service └─24797 /bin/java -Xms4g -Xmx4g -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=75 -XX:+UseCMSInitiatingOccupancyOnly -Djav…
Sep 16 00:28:24 SELKS2 logstash[24797]: at org.jruby.RubyProc.call(RubyProc.java:318) Sep 16 00:28:24 SELKS2 logstash[24797]: at org.jruby.internal.runtime.RubyRunnable.run(RubyRunnable.java:105) Sep 16 00:28:24 SELKS2 logstash[24797]: at java.base/java.lang.Thread.run(Thread.java:834) Sep 16 00:28:24 SELKS2 logstash[24797]: Caused by: java.io.IOException: No space left on device Sep 16 00:28:24 SELKS2 logstash[24797]: at java.base/java.io.FileOutputStream.writeBytes(Native Method) Sep 16 00:28:24 SELKS2 logstash[24797]: at java.base/java.io.FileOutputStream.write(FileOutputStream.java:354) Sep 16 00:28:24 SELKS2 logstash[24797]: at org.apache.logging.log4j.core.appender.OutputStreamManager.writeToDestination(Output….java:250) Sep 16 00:28:24 SELKS2 logstash[24797]: ... 83 more Sep 16 00:28:24 SELKS2 logstash[24797]: [2020-09-16T00:28:24,283][INFO ][logstash.outputs.elasticsearch][main][2ada9d36290a6a5138e7215602be65b629… Sep 16 00:28:24 SELKS2 logstash[24797]: [2020-09-16T00:28:24,286][INFO ][logstash.outputs.elasticsearch][main][2ada9d36290a6a5138e72156…:count=>2} Hint: Some lines were ellipsized, use -l to show in full. ● kibana.service - Kibana Loaded: loaded (/etc/systemd/system/kibana.service; enabled; vendor preset: enabled) Active: active (running) since Wed 2020-09-16 11:14:15 CEST; 1h 17min ago Main PID: 22143 (node) Tasks: 11 (limit: 4915) Memory: 940.2M CGroup: /system.slice/kibana.service └─22143 /usr/share/kibana/bin/../node/bin/node /usr/share/kibana/bin/../src/cli
Sep 16 11:14:51 SELKS2 kibana[22143]: {"type":"log","@timestamp":"2020-09-16T09:14:51Z","tags":["info","savedobjects-service"],"pid":2…igrations"} Sep 16 11:14:51 SELKS2 kibana[22143]: {"type":"log","@timestamp":"2020-09-16T09:14:51Z","tags":["warning","plugins","reporting","config"],"pid":2… Sep 16 11:14:51 SELKS2 kibana[22143]: {"type":"log","@timestamp":"2020-09-16T09:14:51Z","tags":["info","savedobjects-service"],"pid":2…kibana_2."} Sep 16 11:14:51 SELKS2 kibana[22143]: {"type":"log","@timestamp":"2020-09-16T09:14:51Z","tags":["warning","savedobjects-service"],"pid":22143,"me… Sep 16 11:14:51 SELKS2 kibana[22143]: {"type":"log","@timestamp":"2020-09-16T09:14:51Z","tags":["warning","savedobjects-service"],"pid":22143,"me… Sep 16 11:17:38 SELKS2 kibana[22143]: {"type":"log","@timestamp":"2020-09-16T09:17:38Z","tags":["info","savedobjects-service"],"pid":2…espaces\""} Sep 16 11:33:02 SELKS2 kibana[22143]: {"type":"log","@timestamp":"2020-09-16T09:33:02Z","tags":["info","savedobjects-service"],"pid":2…espaces\""} Sep 16 11:33:03 SELKS2 kibana[22143]: {"type":"log","@timestamp":"2020-09-16T09:33:03Z","tags":["info","savedobjects-service"],"pid":2…espaces\""} Sep 16 11:42:22 SELKS2 kibana[22143]: {"type":"log","@timestamp":"2020-09-16T09:42:22Z","tags":["info","savedobjects-service"],"pid":2…espaces\""} Sep 16 11:42:23 SELKS2 kibana[22143]: {"type":"log","@timestamp":"2020-09-16T09:42:23Z","tags":["info","savedobjects-service"],"pid":2…espaces\""} Hint: Some lines were ellipsized, use -l to show in full. ● evebox.service - EveBox Server Loaded: loaded (/lib/systemd/system/evebox.service; enabled; vendor preset: enabled) Active: active (running) since Tue 2020-08-25 15:16:20 CEST; 3 weeks 0 days ago Main PID: 599 (evebox) Tasks: 8 (limit: 4915) Memory: 21.6M CGroup: /system.slice/evebox.service └─599 /usr/bin/evebox server
Warning: Journal has been rotated since unit was started. Log output is incomplete or unavailable. ● molochviewer-selks.service - Moloch Viewer Loaded: loaded (/etc/systemd/system/molochviewer-selks.service; enabled; vendor preset: enabled) Active: active (running) since Wed 2020-09-16 11:14:18 CEST; 1h 17min ago Main PID: 22213 (sh) Tasks: 12 (limit: 4915) Memory: 79.6M CGroup: /system.slice/molochviewer-selks.service ├─22213 /bin/sh -c /data/moloch/bin/node viewer.js -c /data/moloch/etc/config.ini >> /data/moloch/logs/viewer.log 2>&1 └─22220 /data/moloch/bin/node viewer.js -c /data/moloch/etc/config.ini
Sep 16 11:14:18 SELKS2 systemd[1]: Started Moloch Viewer. Sep 16 11:17:20 SELKS2 systemd[1]: molochviewer-selks.service: Current command vanished from the unit file, execution of the command l…be resumed. Hint: Some lines were ellipsized, use -l to show in full. ● molochpcapread-selks.service - Moloch Pcap Read Loaded: loaded (/etc/systemd/system/molochpcapread-selks.service; enabled; vendor preset: enabled) Active: active (running) since Wed 2020-09-16 11:14:18 CEST; 1h 17min ago Main PID: 22209 (sh) Tasks: 6 (limit: 4915) Memory: 185.6M CGroup: /system.slice/molochpcapread-selks.service ├─22209 /bin/sh -c /data/moloch/bin/moloch-capture -c /data/moloch/etc/config.ini -m --copy --delete -R /data/nsm/ >> /data/moloch/lo… └─22212 /data/moloch/bin/moloch-capture -c /data/moloch/etc/config.ini -m --copy --delete -R /data/nsm/
Sep 16 11:14:18 SELKS2 systemd[1]: Started Moloch Pcap Read. Sep 16 11:17:20 SELKS2 systemd[1]: molochpcapread-selks.service: Current command vanished from the unit file, execution of the command…be resumed. Hint: Some lines were ellipsized, use -l to show in full. scirius RUNNING pid 22164, uptime 1:17:53 ii elasticsearch 7.9.1 amd64 Distributed RESTful search engine built for the cloud ii elasticsearch-curator 5.8.1 amd64 Have indices in Elasticsearch? This is the tool for you!\n\nLike a museum curator manages the exhibits and collections on display, \nElasticsearch Curator helps you curate, or manage your indices. ii evebox 1:0.11.1 amd64 no description given ii kibana 7.9.1 amd64 Explore and visualize your Elasticsearch data ii kibana-dashboards-stamus 2020042401 amd64 Kibana 6 dashboard templates. ii logstash 1:7.9.1-1 all An extensible logging pipeline ii moloch 2.4.0-1 amd64 Moloch Full Packet System ii scirius 3.5.0-3 amd64 Django application to manage Suricata ruleset ii suricata 1:2020091301-0stamus0 amd64 Suricata open source multi-thread IDS/IPS/NSM system. Filesystem Type Size Used Avail Use% Mounted on udev devtmpfs 7.8G 0 7.8G 0% /dev tmpfs tmpfs 1.6G 157M 1.5G 10% /run /dev/md0 ext4 887G 519G 324G 62% / tmpfs tmpfs 7.8G 0 7.8G 0% /dev/shm tmpfs tmpfs 5.0M 0 5.0M 0% /run/lock tmpfs tmpfs 7.8G 0 7.8G 0% /sys/fs/cgroup tmpfs tmpfs 1.6G 0 1.6G 0% /run/user/1001
Sep 16 00:28:24 SELKS2 logstash[24797]: Caused by: java.io.IOException: No space left on device
Seems this could be the reason.
Then you need to Clean up old logs from ES, restart ES etc..
I think the problem is in kibana dashboards confguration. Look
START of first time setup script - Wed 16 Sep 2020 11:31:43 AM CEST
Please supply a network interface(s) to set up SELKS Suricata IDPS thread detection on 0: enp0s31f6 1: enp1s0 2: lo Please type in interface or space delimited interfaces below and hit "Enter". Example: eth1 OR Example: eth1 eth2 eth3
Configure threat detection for INTERFACE(S):
The supplied network interface(s): enp0s31f6
DONE! FPC - Full Packet Capture. Suricata will rotate and delete the pcap captured files. FPC_Retain - Full Packet Capture with having Moloch's pcap retention/rotation. Keeps the pcaps as long as there is space available. None - disable packet capture
1) FPC 2) FPC_Retain 3) NONE Please choose an option. Type in a number and hit "Enter" Enable Full Pcacket Capture with pcap retaining
% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 404 100 404 0 0 394k 0 --:--:-- --:--:-- --:--:-- 394k {"cluster_name":"elasticsearch","status":"yellow","timed_out":false,"number_of_nodes":1,"number_of_data_nodes":1,"active_primary_shards":91,"active_shards":91,"relocating_shards":0,"initializing_shards":0,"unassigned_shards":2,"delayed_unassigned_shards":0,"number_of_pending_tasks":0,"number_of_in_flight_fetch":0,"task_max_waiting_in_queue_millis":0,"active_shards_percent_as_number":97.84946236559139}
WARNING elasticsearch health is 'yellow' instead of 'green', things may be broken
It is STRONGLY recommended that you stop ALL moloch captures and viewers before proceeding. Use 'db.pl http://localhost:9200 backup' to backup db first.
There is 1 elastic search data node, if you expect more please fix first before proceeding.
It appears this elastic search cluster already has moloch installed (version 64), this will delete ALL data in elastic search! (It does not delete the pcap files on disk.)
Type "INIT" to continue - do you want to erase everything?? Erasing Creating
Finished Found interfaces: enp0s31f6;enp1s0;lo Semicolon ';' seperated list of interfaces to monitor [eth1] Install Elasticsearch server locally for demo, must have at least 3G of memory, NOT recommended for production use (yes or no) [no] Elasticsearch server URL [http://localhost:9200] Password to encrypt S2S and other things [no-default] Moloch - Creating configuration files Not overwriting /data/moloch/etc/config.ini, delete and run again if update required (usually not), or edit by hand Installing systemd start files, use systemctl Download GEO files? (yes or no) [yes] Moloch - Downloading GEO files 2020-09-16 11:32:27 URL:https://raw.githubusercontent.com/wireshark/wireshark/master/manuf [1750878/1750878] -> "oui.txt" [1]
Moloch - Configured - Now continue with step 4 in /data/moloch/README.txt
/sbin/start elasticsearch # for upstart/Centos 6/Ubuntu 14.04
systemctl start elasticsearch.service # for systemd/Centos 7/Ubuntu 16.04
5) Initialize/Upgrade Elasticsearch Moloch configuration a) If this is the first install, or want to delete all data /data/moloch/db/db.pl http://ESHOST:9200 init b) If this is an update to moloch package /data/moloch/db/db.pl http://ESHOST:9200 upgrade 6) Add an admin user if a new install or after an init /data/moloch/bin/moloch_add_user.sh admin "Admin User" THEPASSWORD --admin 7) Start everything a) If using upstart (Centos 6 or sometimes Ubuntu 14.04): /sbin/start molochcapture /sbin/start molochviewer b) If using systemd (Centos 7 or Ubuntu 16.04 or sometimes Ubuntu 14.04) systemctl start molochcapture.service systemctl start molochviewer.service 8) Look at log files for errors /data/moloch/logs/viewer.log /data/moloch/logs/capture.log 9) Visit http://MOLOCHHOST:8005 with your favorite browser. user: admin password: THEPASSWORD from step #6
If you want IP -> Geo/ASN to work, you need to setup a maxmind account and the geoipupdate program. See https://molo.ch/faq#maxmind
Any configuration changes can be made to /data/moloch/etc/config.ini See https://molo.ch/faq#moloch-is-not-working for issues
Additional information can be found at:
Would you like to setup a retention policy now? (y/n)
Please specify the maximum file size in Gigabytes. The disk should have room for at least 10 times the specified value. (default is 12)
Setting maxFileSizeG to 15 Gigabyte.
Please specify the maximum rotation time in minutes. (default is none)
Setting maxFileTimeM to 600 minutes.
Added
Traceback (most recent call last):
File "bin/manage.py", line 10, in
FINISH of first time setup script - Wed 16 Sep 2020 11:33:22 AM CEST
Exited with FAILED Full log located at - /opt/selks/log/selks-first-time-setup_stamus.log Press enter to continue
Is it possible to try to solve the full disk issue first to remove it as a point?
Problem is - disk is not full.
/dev/md0 ext4 887G 519G 324G 62% /
The disk has 324GB free from any 887GB.
So, the logstash Error about disk full is any false positive message.
I tried to reboot the SELKS device, but now is totally dead (grub_calloc not found error). So, first I will repair the grub error and restart the device.
I repaired the grub data and rebooted the SELKS device.
Still the same error.
Hard disk have space enough (835 GB)
scirius RUNNING pid 857, uptime 0:18:48
ii elasticsearch 7.9.1 amd64 Distributed RESTful search engine built for the cloud
ii elasticsearch-curator 5.8.1 amd64 Have indices in Elasticsearch? This is the tool for you!\n\nLike a museum curator manages the exhibits and collections on display, \nElasticsearch Curator helps you curate, or manage your indices.
ii evebox 1:0.11.1 amd64 no description given
ii kibana 7.9.1 amd64 Explore and visualize your Elasticsearch data
ii kibana-dashboards-stamus 2020042401 amd64 Kibana 6 dashboard templates.
ii logstash 1:7.9.1-1 all An extensible logging pipeline
ii moloch 2.4.0-1 amd64 Moloch Full Packet System
ii scirius 3.5.0-3 amd64 Django application to manage Suricata ruleset
ii suricata 1:2020091301-0stamus0 amd64 Suricata open source multi-thread IDS/IPS/NSM system.
Filesystem Type Size Used Avail Use% Mounted on
udev devtmpfs 7.8G 0 7.8G 0% /dev
tmpfs tmpfs 1.6G 8.9M 1.6G 1% /run
/dev/md0 ext4 887G 6.9G 835G 1% /
tmpfs tmpfs 7.8G 0 7.8G 0% /dev/shm
tmpfs tmpfs 5.0M 0 5.0M 0% /run/lock
tmpfs tmpfs 7.8G 0 7.8G 0% /sys/fs/cgroup
tmpfs tmpfs 1.6G 0 1.6G 0% /run/user/1001
Any idea?
Can you please share the last 100 lines form the logs of elasticsearch and logstash ?
I am trying to figure out if this is not an issue - https://stackoverflow.com/questions/34911181/how-to-undo-setting-elasticsearch-index-to-readonly
Here is elasticsearch
root@SELKS2:/var/log/elasticsearch# tail -100 elasticsearch.log [2020-09-27T02:00:01,333][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T02:00:01,334][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T02:00:01,404][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-flow-2020.09.27/yAy9Z-_IQhaAbhMCq9BU5w] update_mapping [_doc] [2020-09-27T02:00:01,481][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-flow-2020.09.27/yAy9Z-_IQhaAbhMCq9BU5w] update_mapping [_doc] [2020-09-27T02:00:01,766][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-fileinfo-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:00:01,846][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-tls-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:00:02,032][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-http-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:00:02,181][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-alert-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:00:02,555][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-flow-2020.09.27/yAy9Z-_IQhaAbhMCq9BU5w] update_mapping [_doc] [2020-09-27T02:00:02,739][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-tls-2020.09.27/xVw_m4z2RruVAEmFt6sZGQ] update_mapping [_doc] [2020-09-27T02:00:02,747][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-fileinfo-2020.09.27/lSUFo6xuS2-QbBCPAwpTnw] update_mapping [_doc] [2020-09-27T02:00:02,755][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:00:02,765][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T02:00:02,856][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-alert-2020.09.27/ZEcq-MPcRiusFFv0Q7DeYw] update_mapping [_doc] [2020-09-27T02:00:02,968][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T02:00:03,079][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:00:03,228][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:00:03,236][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-fileinfo-2020.09.27/lSUFo6xuS2-QbBCPAwpTnw] update_mapping [_doc] [2020-09-27T02:00:03,524][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-2020.09.27/wU-tIVRJRpS0vsqMThTQ3g] update_mapping [_doc] [2020-09-27T02:00:03,780][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-tls-2020.09.27/xVw_m4z2RruVAEmFt6sZGQ] update_mapping [_doc] [2020-09-27T02:00:04,781][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-tls-2020.09.27/xVw_m4z2RruVAEmFt6sZGQ] update_mapping [_doc] [2020-09-27T02:00:04,783][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-alert-2020.09.27/ZEcq-MPcRiusFFv0Q7DeYw] update_mapping [_doc] [2020-09-27T02:00:04,873][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T02:00:04,880][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T02:00:05,790][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-dhcp-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:00:05,901][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-anomaly-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:00:06,062][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-smtp-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:00:06,356][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-fileinfo-2020.09.27/lSUFo6xuS2-QbBCPAwpTnw] update_mapping [_doc] [2020-09-27T02:00:06,366][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:00:06,538][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dhcp-2020.09.27/v3c2212RSJmSxcWrs_jn-w] update_mapping [_doc] [2020-09-27T02:00:06,558][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-anomaly-2020.09.27/3Zdi4g4AT2WDfRadG47HVA] update_mapping [_doc] [2020-09-27T02:00:06,632][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-fileinfo-2020.09.27/lSUFo6xuS2-QbBCPAwpTnw] update_mapping [_doc] [2020-09-27T02:00:06,638][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T02:00:06,643][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:00:06,648][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-smtp-2020.09.27/JeJcc0IcRICMWQq3oQNMwQ] update_mapping [_doc] [2020-09-27T02:00:06,835][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-flow-2020.09.27/yAy9Z-_IQhaAbhMCq9BU5w] update_mapping [_doc] [2020-09-27T02:00:08,793][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-alert-2020.09.27/ZEcq-MPcRiusFFv0Q7DeYw] update_mapping [_doc] [2020-09-27T02:00:09,803][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-snmp-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:00:09,927][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T02:00:10,172][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-snmp-2020.09.27/z7fQ7mXJSvqYVeoF0duODw] update_mapping [_doc] [2020-09-27T02:00:10,261][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-snmp-2020.09.27/z7fQ7mXJSvqYVeoF0duODw] update_mapping [_doc] [2020-09-27T02:00:14,821][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T02:00:20,241][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-2020.09.27/wU-tIVRJRpS0vsqMThTQ3g] update_mapping [_doc] [2020-09-27T02:00:20,830][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T02:00:22,839][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:00:41,710][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [sessions2-200927] creating index, cause [auto(bulk api)], templates [sessions2_template], shards [1]/[0] [2020-09-27T02:00:45,992][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-flow-2020.09.27/yAy9Z-_IQhaAbhMCq9BU5w] update_mapping [_doc] [2020-09-27T02:01:14,439][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:02:12,521][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-snmp-2020.09.27/z7fQ7mXJSvqYVeoF0duODw] update_mapping [_doc] [2020-09-27T02:02:12,610][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-alert-2020.09.27/ZEcq-MPcRiusFFv0Q7DeYw] update_mapping [_doc] [2020-09-27T02:04:42,777][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:04:43,775][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-fileinfo-2020.09.27/lSUFo6xuS2-QbBCPAwpTnw] update_mapping [_doc] [2020-09-27T02:10:47,516][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:10:47,588][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-fileinfo-2020.09.27/lSUFo6xuS2-QbBCPAwpTnw] update_mapping [_doc] [2020-09-27T02:14:05,969][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:15:42,159][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:15:42,243][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-fileinfo-2020.09.27/lSUFo6xuS2-QbBCPAwpTnw] update_mapping [_doc] [2020-09-27T02:15:43,161][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T02:18:05,450][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-tftp-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:18:05,784][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-tftp-2020.09.27/unaBzackSDm9GgVJ2owAkQ] update_mapping [_doc] [2020-09-27T02:18:43,548][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-sip-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:18:43,913][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-sip-2020.09.27/T1wV1HRzTxOssSarhOh1cQ] update_mapping [_doc] [2020-09-27T02:21:50,058][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:24:05,304][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-smtp-2020.09.27/JeJcc0IcRICMWQq3oQNMwQ] update_mapping [_doc] [2020-09-27T02:27:20,807][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:27:20,904][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-fileinfo-2020.09.27/lSUFo6xuS2-QbBCPAwpTnw] update_mapping [_doc] [2020-09-27T02:27:44,864][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:35:33,042][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:35:45,073][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:41:16,848][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-snmp-2020.09.27/z7fQ7mXJSvqYVeoF0duODw] update_mapping [_doc] [2020-09-27T02:42:38,033][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-sip-2020.09.27/T1wV1HRzTxOssSarhOh1cQ] update_mapping [_doc] [2020-09-27T02:44:49,385][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-alert-2020.09.27/ZEcq-MPcRiusFFv0Q7DeYw] update_mapping [_doc] [2020-09-27T02:44:49,510][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-tls-2020.09.27/xVw_m4z2RruVAEmFt6sZGQ] update_mapping [_doc] [2020-09-27T02:47:58,821][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-flow-2020.09.27/yAy9Z-_IQhaAbhMCq9BU5w] update_mapping [_doc] [2020-09-27T02:48:07,842][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-anomaly-2020.09.27/3Zdi4g4AT2WDfRadG47HVA] update_mapping [_doc] [2020-09-27T02:50:33,179][INFO ][o.e.c.m.MetadataCreateIndexService] [SELKS2] [logstash-ssh-2020.09.27] creating index, cause [auto(bulk api)], templates [logstash], shards [1]/[0] [2020-09-27T02:50:33,484][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-ssh-2020.09.27/RGrYvMxyQJefzo8NLPdRAA] update_mapping [_doc] [2020-09-27T02:52:31,509][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T02:57:34,159][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-dns-2020.09.27/15rXpPC4Q42ArXPxkhPVcw] update_mapping [_doc] [2020-09-27T03:13:26,433][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T03:23:02,802][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-flow-2020.09.27/yAy9Z-_IQhaAbhMCq9BU5w] update_mapping [_doc] [2020-09-27T03:30:00,001][INFO ][o.e.x.s.SnapshotRetentionTask] [SELKS2] starting SLM retention snapshot cleanup task [2020-09-27T03:30:00,002][INFO ][o.e.x.s.SnapshotRetentionTask] [SELKS2] there are no repositories to fetch, SLM retention snapshot cleanup task complete [2020-09-27T03:52:26,854][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-alert-2020.09.27/ZEcq-MPcRiusFFv0Q7DeYw] update_mapping [_doc] [2020-09-27T04:06:49,923][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-alert-2020.09.27/ZEcq-MPcRiusFFv0Q7DeYw] update_mapping [_doc] [2020-09-27T04:06:49,927][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-tls-2020.09.27/xVw_m4z2RruVAEmFt6sZGQ] update_mapping [_doc] [2020-09-27T04:12:16,691][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-fileinfo-2020.09.27/lSUFo6xuS2-QbBCPAwpTnw] update_mapping [_doc] [2020-09-27T04:12:16,814][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T04:36:24,034][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-sip-2020.09.27/T1wV1HRzTxOssSarhOh1cQ] update_mapping [_doc] [2020-09-27T05:22:47,920][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-alert-2020.09.27/ZEcq-MPcRiusFFv0Q7DeYw] update_mapping [_doc] [2020-09-27T05:45:40,046][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-fileinfo-2020.09.27/lSUFo6xuS2-QbBCPAwpTnw] update_mapping [_doc] [2020-09-27T05:45:40,116][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T06:18:47,834][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc] [2020-09-27T06:18:47,941][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-fileinfo-2020.09.27/lSUFo6xuS2-QbBCPAwpTnw] update_mapping [_doc] [2020-09-27T06:20:05,047][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-ssh-2020.09.27/RGrYvMxyQJefzo8NLPdRAA] update_mapping [_doc] [2020-09-27T07:44:37,021][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-smtp-2020.09.27/JeJcc0IcRICMWQq3oQNMwQ] update_mapping [_doc] [2020-09-27T08:35:28,158][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-alert-2020.09.27/ZEcq-MPcRiusFFv0Q7DeYw] update_mapping [_doc] [2020-09-27T09:20:33,707][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-alert-2020.09.27/ZEcq-MPcRiusFFv0Q7DeYw] update_mapping [_doc] [2020-09-27T09:22:11,908][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-alert-2020.09.27/ZEcq-MPcRiusFFv0Q7DeYw] update_mapping [_doc] [2020-09-27T10:13:10,514][INFO ][o.e.c.m.MetadataMappingService] [SELKS2] [logstash-http-2020.09.27/RqUmP8jrQQiD1rnqsowXMQ] update_mapping [_doc]
And logstash
root@SELKS2:/var/log/logstash# tail -100 logstash-plain.log
[2020-09-23T12:27:03,899][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:27:05,513][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:27:08,908][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:27:10,557][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:27:13,915][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:27:15,601][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:27:18,924][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:27:20,645][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:27:23,931][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:27:25,689][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:27:28,939][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:27:30,733][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:27:33,946][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:27:35,779][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:27:38,954][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:27:40,823][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:27:43,961][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:27:45,867][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:27:48,968][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:27:50,911][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:27:53,979][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:27:55,955][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:27:58,988][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:00,999][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:02,786][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c03be43b13] Attempted to send a bulk request to elasticsearch, but no there are no living connections in the connection pool. Perhaps Elasticsearch is unreachable or down? {:error_message=>"No Available connections", :class=>"LogStash::Outputs::ElasticSearch::HttpClient::Pool::NoConnectionAvailableError", :will_retry_in_seconds=>64}
[2020-09-23T12:28:02,787][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c03be43b13] Attempted to send a bulk request to elasticsearch, but no there are no living connections in the connection pool. Perhaps Elasticsearch is unreachable or down? {:error_message=>"No Available connections", :class=>"LogStash::Outputs::ElasticSearch::HttpClient::Pool::NoConnectionAvailableError", :will_retry_in_seconds=>64}
[2020-09-23T12:28:02,788][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c03be43b13] Attempted to send a bulk request to elasticsearch, but no there are no living connections in the connection pool. Perhaps Elasticsearch is unreachable or down? {:error_message=>"No Available connections", :class=>"LogStash::Outputs::ElasticSearch::HttpClient::Pool::NoConnectionAvailableError", :will_retry_in_seconds=>64}
[2020-09-23T12:28:02,791][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c03be43b13] Attempted to send a bulk request to elasticsearch, but no there are no living connections in the connection pool. Perhaps Elasticsearch is unreachable or down? {:error_message=>"No Available connections", :class=>"LogStash::Outputs::ElasticSearch::HttpClient::Pool::NoConnectionAvailableError", :will_retry_in_seconds=>64}
[2020-09-23T12:28:03,995][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:06,038][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:09,002][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:11,079][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:14,009][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:16,118][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:19,016][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:21,163][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:24,023][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:26,207][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:29,030][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:31,252][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:34,040][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:36,296][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:39,046][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:41,340][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:44,053][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:46,384][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:49,060][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:51,428][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:54,067][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:28:56,472][WARN ][org.logstash.execution.ShutdownWatcherExt] {"inflight_count"=>0, "stalling_threads_info"=>{["LogStash::Filters::Metrics", {"add_tag"=>"metric", "id"=>"c0db4298d0ed6050a2f6b8477f5838e46c80764e15c899706176877a209062a8", "flush_interval"=>30, "meter"=>["eve_insert"]}]=>[{"thread_id"=>32, "name"=>"[main]>worker0", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>33, "name"=>"[main]>worker1", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}, {"thread_id"=>34, "name"=>"[main]>worker2", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in sleep'"}, {"thread_id"=>35, "name"=>"[main]>worker3", "current_call"=>"[...]/vendor/bundle/jruby/2.5.0/gems/stud-0.0.23/lib/stud/interval.rb:89:in
sleep'"}]}}
[2020-09-23T12:28:59,074][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, but got an error. {:url=>"http://127.0.0.1:9200/", :error_type=>LogStash::Outputs::ElasticSearch::HttpClient::Pool::HostUnreachableError, :error=>"Elasticsearch Unreachable: [http://127.0.0.1:9200/][Manticore::SocketException] Connection refused (Connection refused)"}
[2020-09-23T12:34:55,260][INFO ][logstash.runner ] Starting Logstash {"logstash.version"=>"7.9.1", "jruby.version"=>"jruby 9.2.13.0 (2.5.7) 2020-08-03 9a89c94bcc OpenJDK 64-Bit Server VM 11.0.8+10-post-Debian-1deb10u1 on 11.0.8+10-post-Debian-1deb10u1 +indy +jit [linux-x86_64]"}
[2020-09-23T12:34:57,575][INFO ][org.reflections.Reflections] Reflections took 28 ms to scan 1 urls, producing 22 keys and 45 values
[2020-09-23T12:34:58,322][INFO ][logstash.outputs.elasticsearch][main] Elasticsearch pool URLs updated {:changes=>{:removed=>[], :added=>[http://127.0.0.1:9200/]}}
[2020-09-23T12:34:58,480][WARN ][logstash.outputs.elasticsearch][main] Restored connection to ES instance {:url=>"http://127.0.0.1:9200/"}
[2020-09-23T12:34:58,524][INFO ][logstash.outputs.elasticsearch][main] ES Output version determined {:es_version=>7}
[2020-09-23T12:34:58,528][WARN ][logstash.outputs.elasticsearch][main] Detected a 6.x and above cluster: the type
event field won't be used to determine the document _type {:es_version=>7}
[2020-09-23T12:34:58,588][INFO ][logstash.outputs.elasticsearch][main] New Elasticsearch output {:class=>"LogStash::Outputs::ElasticSearch", :hosts=>["//127.0.0.1"]}
[2020-09-23T12:34:58,598][INFO ][logstash.outputs.elasticsearch][main] Elasticsearch pool URLs updated {:changes=>{:removed=>[], :added=>[http://127.0.0.1:9200/]}}
[2020-09-23T12:34:58,607][WARN ][logstash.outputs.elasticsearch][main] Restored connection to ES instance {:url=>"http://127.0.0.1:9200/"}
[2020-09-23T12:34:58,614][INFO ][logstash.outputs.elasticsearch][main] ES Output version determined {:es_version=>7}
[2020-09-23T12:34:58,621][WARN ][logstash.outputs.elasticsearch][main] Detected a 6.x and above cluster: the type
event field won't be used to determine the document _type {:es_version=>7}
[2020-09-23T12:34:58,637][INFO ][logstash.outputs.elasticsearch][main] Using mapping template from {:path=>"/etc/logstash/elasticsearch7-template.json"}
[2020-09-23T12:34:58,659][INFO ][logstash.outputs.elasticsearch][main] New Elasticsearch output {:class=>"LogStash::Outputs::ElasticSearch", :hosts=>["//127.0.0.1"]}
[2020-09-23T12:34:58,715][INFO ][logstash.outputs.elasticsearch][main] Attempting to install template {:manage_template=>{"template"=>"logstash-", "version"=>60001, "settings"=>{"number_of_replicas"=>0, "index.refresh_interval"=>"5s"}, "mappings"=>{"dynamic_templates"=>[{"message_field"=>{"path_match"=>"message", "match_mapping_type"=>"string", "mapping"=>{"type"=>"text", "norms"=>false}}}, {"string_fields"=>{"match"=>"", "match_mapping_type"=>"string", "mapping"=>{"type"=>"text", "norms"=>false, "fields"=>{"keyword"=>{"type"=>"keyword", "ignore_above"=>256}}}}}], "properties"=>{"@timestamp"=>{"type"=>"date"}, "@version"=>{"type"=>"keyword"}, "geoip"=>{"dynamic"=>true, "properties"=>{"ip"=>{"type"=>"ip"}, "location"=>{"type"=>"geo_point"}, "latitude"=>{"type"=>"half_float"}, "longitude"=>{"type"=>"half_float"}}}}}}}
[2020-09-23T12:34:58,683][INFO ][logstash.outputs.elasticsearch][main] Using mapping template from {:path=>"/etc/logstash/elasticsearch7-template.json"}
[2020-09-23T12:34:58,749][INFO ][logstash.outputs.elasticsearch][main] Attempting to install template {:manage_template=>{"template"=>"logstash-", "version"=>60001, "settings"=>{"number_of_replicas"=>0, "index.refresh_interval"=>"5s"}, "mappings"=>{"dynamic_templates"=>[{"message_field"=>{"path_match"=>"message", "match_mapping_type"=>"string", "mapping"=>{"type"=>"text", "norms"=>false}}}, {"string_fields"=>{"match"=>"", "match_mapping_type"=>"string", "mapping"=>{"type"=>"text", "norms"=>false, "fields"=>{"keyword"=>{"type"=>"keyword", "ignore_above"=>256}}}}}], "properties"=>{"@timestamp"=>{"type"=>"date"}, "@version"=>{"type"=>"keyword"}, "geoip"=>{"dynamic"=>true, "properties"=>{"ip"=>{"type"=>"ip"}, "location"=>{"type"=>"geo_point"}, "latitude"=>{"type"=>"half_float"}, "longitude"=>{"type"=>"half_float"}}}}}}}
[2020-09-23T12:34:58,755][INFO ][logstash.outputs.elasticsearch][main] Installing elasticsearch template to _template/logstash
[2020-09-23T12:34:58,758][INFO ][logstash.outputs.elasticsearch][main] Installing elasticsearch template to _template/logstash
[2020-09-23T12:34:58,912][INFO ][logstash.filters.geoip ][main] Using geoip database {:path=>"/usr/share/logstash/vendor/bundle/jruby/2.5.0/gems/logstash-filter-geoip-6.0.3-java/vendor/GeoLite2-City.mmdb"}
[2020-09-23T12:34:59,029][INFO ][logstash.filters.geoip ][main] Using geoip database {:path=>"/usr/share/logstash/vendor/bundle/jruby/2.5.0/gems/logstash-filter-geoip-6.0.3-java/vendor/GeoLite2-City.mmdb"}
[2020-09-23T12:34:59,082][INFO ][logstash.javapipeline ][main] Starting pipeline {:pipeline_id=>"main", "pipeline.workers"=>4, "pipeline.batch.size"=>125, "pipeline.batch.delay"=>50, "pipeline.max_inflight"=>500, "pipeline.sources"=>["/etc/logstash/conf.d/logstash.conf", "/etc/logstash/conf.d/scirius-logstash.conf"], :thread=>"#
And health root@SELKS2:/var/log/logstash# selks-health-check_stamus ● suricata.service - LSB: Next Generation IDS/IPS Loaded: loaded (/etc/init.d/suricata; generated) Active: active (running) since Sun 2020-09-27 02:00:44 CEST; 8h ago Docs: man:systemd-sysv-generator(8) Process: 13131 ExecStart=/etc/init.d/suricata start (code=exited, status=0/SUCCESS) Tasks: 10 (limit: 4915) Memory: 821.1M CGroup: /system.slice/suricata.service └─13139 /usr/bin/suricata -c /etc/suricata/suricata.yaml --pidfile /var/run/suricata.pid --af-packet -D -v --user=logstash
Sep 27 02:00:44 SELKS2 systemd[1]: Starting LSB: Next Generation IDS/IPS... Sep 27 02:00:44 SELKS2 suricata[13131]: Starting suricata in IDS (af-packet) mode... done. Sep 27 02:00:44 SELKS2 systemd[1]: Started LSB: Next Generation IDS/IPS. ● elasticsearch.service - Elasticsearch Loaded: loaded (/lib/systemd/system/elasticsearch.service; enabled; vendor preset: enabled) Active: active (running) since Wed 2020-09-23 12:53:48 CEST; 3 days ago Docs: https://www.elastic.co Main PID: 4392 (java) Tasks: 94 (limit: 4915) Memory: 9.4G CGroup: /system.slice/elasticsearch.service ├─4392 /usr/share/elasticsearch/jdk/bin/java -Xshare:auto -Des.networkaddress.cache.ttl=60 -Des.networkaddress.cache.negative.ttl=10 -XX:+AlwaysPreTouch … └─4575 /usr/share/elasticsearch/modules/x-pack-ml/platform/linux-x86_64/bin/controller
Sep 23 12:53:39 SELKS2 systemd[1]: Starting Elasticsearch... Sep 23 12:53:48 SELKS2 systemd[1]: Started Elasticsearch. ● logstash.service - logstash Loaded: loaded (/etc/systemd/system/logstash.service; enabled; vendor preset: enabled) Active: active (running) since Wed 2020-09-23 12:34:21 CEST; 3 days ago Main PID: 635 (java) Tasks: 41 (limit: 4915) Memory: 1.1G CGroup: /system.slice/logstash.service └─635 /bin/java -Xms4g -Xmx4g -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=75 -XX:+UseCMSInitiatingOccupancyOnly -Djava.awt.headless=true -…
Sep 23 12:53:45 SELKS2 logstash[635]: [2020-09-23T12:53:45,431][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c… Sep 23 12:53:45 SELKS2 logstash[635]: [2020-09-23T12:53:45,430][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c… Sep 23 12:53:45 SELKS2 logstash[635]: [2020-09-23T12:53:45,431][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c… Sep 23 12:53:45 SELKS2 logstash[635]: [2020-09-23T12:53:45,430][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c… Sep 23 12:53:48 SELKS2 logstash[635]: [2020-09-23T12:53:48,900][WARN ][logstash.outputs.elasticsearch][main] Attempted to resurrect connection to dead ES instance, … Sep 23 12:53:53 SELKS2 logstash[635]: [2020-09-23T12:53:53,443][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c… Sep 23 12:53:53 SELKS2 logstash[635]: [2020-09-23T12:53:53,443][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c… Sep 23 12:53:53 SELKS2 logstash[635]: [2020-09-23T12:53:53,444][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c… Sep 23 12:53:53 SELKS2 logstash[635]: [2020-09-23T12:53:53,445][ERROR][logstash.outputs.elasticsearch][main][e55f734d663b7fb7ca21a05c69227f334d0c6198948f303fac6e50c… Sep 23 12:53:53 SELKS2 logstash[635]: [2020-09-23T12:53:53,956][WARN ][logstash.outputs.elasticsearch][main] Restored connection to ES instance {:url=>…0.0.1:9200/"} Hint: Some lines were ellipsized, use -l to show in full. ● kibana.service - Kibana Loaded: loaded (/etc/systemd/system/kibana.service; enabled; vendor preset: enabled) Active: active (running) since Wed 2020-09-23 12:53:48 CEST; 3 days ago Main PID: 4621 (node) Tasks: 11 (limit: 4915) Memory: 483.3M CGroup: /system.slice/kibana.service └─4621 /usr/share/kibana/bin/../node/bin/node /usr/share/kibana/bin/../src/cli
Sep 23 12:54:01 SELKS2 kibana[4621]: {"type":"log","@timestamp":"2020-09-23T10:54:01Z","tags":["warning","plugins","ingestManager"],"pid":4621,"message…kibana.yml."} Sep 23 12:54:01 SELKS2 kibana[4621]: {"type":"log","@timestamp":"2020-09-23T10:54:01Z","tags":["warning","plugins","actions","actions"],"pid":4621,"mes…kibana.yml."} Sep 23 12:54:01 SELKS2 kibana[4621]: {"type":"log","@timestamp":"2020-09-23T10:54:01Z","tags":["warning","plugins","alerts","plugins","alerting"],"pid"…kibana.yml."} Sep 23 12:54:01 SELKS2 kibana[4621]: {"type":"log","@timestamp":"2020-09-23T10:54:01Z","tags":["info","plugins","monitoring","monitoring"],"pid":4621,"…ion cluster"} Sep 23 12:54:01 SELKS2 kibana[4621]: {"type":"log","@timestamp":"2020-09-23T10:54:01Z","tags":["info","savedobjects-service"],"pid":4621,"message":"Wai…grations..."} Sep 23 12:54:01 SELKS2 kibana[4621]: {"type":"log","@timestamp":"2020-09-23T10:54:01Z","tags":["warning","plugins","reporting","config"],"pid":4621,"message":"Chrom… Sep 23 12:54:01 SELKS2 kibana[4621]: {"type":"log","@timestamp":"2020-09-23T10:54:01Z","tags":["info","savedobjects-service"],"pid":4621,"message":"Sta… migrations"} Sep 23 12:54:01 SELKS2 kibana[4621]: {"type":"log","@timestamp":"2020-09-23T10:54:01Z","tags":["info","savedobjects-service"],"pid":4621,"message":"Cre… .kibana_2."} Sep 23 12:54:01 SELKS2 kibana[4621]: {"type":"log","@timestamp":"2020-09-23T10:54:01Z","tags":["warning","savedobjects-service"],"pid":4621,"message":"Unable to con… Sep 23 12:54:01 SELKS2 kibana[4621]: {"type":"log","@timestamp":"2020-09-23T10:54:01Z","tags":["warning","savedobjects-service"],"pid":4621,"message":"Another Kiban… Hint: Some lines were ellipsized, use -l to show in full. ● evebox.service - EveBox Server Loaded: loaded (/lib/systemd/system/evebox.service; enabled; vendor preset: enabled) Active: active (running) since Wed 2020-09-23 12:34:21 CEST; 3 days ago Main PID: 625 (evebox) Tasks: 8 (limit: 4915) Memory: 40.1M CGroup: /system.slice/evebox.service └─625 /usr/bin/evebox server
Sep 23 12:34:24 SELKS2 evebox[625]: 2020-09-23 12:34:24 (server.go:335)
Sep 23 12:53:51 SELKS2 systemd[1]: Started Moloch Viewer. ● molochpcapread-selks.service - Moloch Pcap Read Loaded: loaded (/etc/systemd/system/molochpcapread-selks.service; enabled; vendor preset: enabled) Active: active (running) since Wed 2020-09-23 12:53:51 CEST; 3 days ago Main PID: 4684 (sh) Tasks: 6 (limit: 4915) Memory: 271.7M CGroup: /system.slice/molochpcapread-selks.service ├─4684 /bin/sh -c /data/moloch/bin/moloch-capture -c /data/moloch/etc/config.ini -m --copy --delete -R /data/nsm/ >> /data/moloch/logs/capture.log 2>&1 └─4686 /data/moloch/bin/moloch-capture -c /data/moloch/etc/config.ini -m --copy --delete -R /data/nsm/
Sep 23 12:53:51 SELKS2 systemd[1]: Started Moloch Pcap Read. scirius RUNNING pid 4640, uptime 3 days, 21:51:31 ii elasticsearch 7.9.1 amd64 Distributed RESTful search engine built for the cloud ii elasticsearch-curator 5.8.1 amd64 Have indices in Elasticsearch? This is the tool for you!\n\nLike a museum curator manages the exhibits and collections on display, \nElasticsearch Curator helps you curate, or manage your indices. ii evebox 1:0.11.1 amd64 no description given ii kibana 7.9.1 amd64 Explore and visualize your Elasticsearch data ii kibana-dashboards-stamus 2020042401 amd64 Kibana 6 dashboard templates. ii logstash 1:7.9.1-1 all An extensible logging pipeline ii moloch 2.4.0-1 amd64 Moloch Full Packet System ii scirius 3.5.0-3 amd64 Django application to manage Suricata ruleset ii suricata 1:2020091301-0stamus0 amd64 Suricata open source multi-thread IDS/IPS/NSM system. Filesystem Type Size Used Avail Use% Mounted on udev devtmpfs 7.8G 0 7.8G 0% /dev tmpfs tmpfs 1.6G 8.9M 1.6G 1% /run /dev/md0 ext4 887G 29G 814G 4% / tmpfs tmpfs 7.8G 0 7.8G 0% /dev/shm tmpfs tmpfs 5.0M 0 5.0M 0% /run/lock tmpfs tmpfs 7.8G 0 7.8G 0% /sys/fs/cgroup tmpfs tmpfs 1.6G 0 1.6G 0% /run/user/1001
The Kibana state is the same
Kibana server is not ready yet
Can you try the following:
?
Done
Kibana seems to be working now.
Elasticsearch is in yellow state. I cleared all data with the script selks-db-logs-cleanup_stamus
Kibana is working, but elasticsearfh has empty window now.
I will wait some hours and report the SELKS state here again.
Dne 29. 09. 20 v 23:34 Peter Manev napsal(a):
Can you try the following:
- stop kibana
- curl -XDELETE http://localhost:9200/.kibana*
- start kibana
- reset/import the dashboards again
?
Now, loooks the problem is passed to logstash x elasticsearch, but Kibana works.
That seems more localized. After you have had some traffic - can you open ok all dashboards in kibana?
It looks like some indexes problem, too.
For that par ti think you need to rerun the moloch setup - as you deleted all indexes/data.
selks-first-time-setup_stamus selks-db-logs-cleanup_stamus
I can see first moloch records. Seems to be working.
Kibana works, Moloch works, but Elastic search gives empty screen, when start search from Kibana
How to regenerate elastic search indexes?
I think you can reset the dashboards now - that will regenerate the indexes. Or simply import the dashboards - like so - https://github.com/StamusNetworks/KTS7#how-to-use
Done, but no effect.
Then i think you need to check inside the Kibana , management - if you have setup a default index (should be logstash-*
) ?
Is it set to logstash-*
ok , so that part is good then. If you open a Kibana dashboard SN-ALERTS - would it populate ok ?
Yes, this dashboard looks good.
in that specific case i think you may refer here - https://github.com/StamusNetworks/SELKS/issues/255
Well, the https://github.com/StamusNetworks/SELKS/issues/255 is not working on my case.
I tried to update the SELKS device with selks-upgrade_stamus
and repair the selks.conf file with https://github.com/StamusNetworks/SELKS/wiki/Kibana-did-not-load-properly
But the result is exactly the same as before the upgrade. Means https://github.com/StamusNetworks/SELKS/issues/251#issuecomment-702052403
What exactly I have to do in https://github.com/StamusNetworks/SELKS/issues/255#issuecomment-698755156 ?
Ok, just so i understand properly - all dashboards/visualizations open properly except these two cases. How do you reproduce that ? Can you share exactly the steps you do to get to the err page - so I can try to reproduce. Pretty please :)
I'm afraid that only way to reproduce the problem is to update an earlier SELKS6 to newer version and keep all config files (so, no update the config files). Means, start again this task https://github.com/StamusNetworks/SELKS/issues/251 . Is it here some way to repair https://github.com/StamusNetworks/SELKS/issues/251#issuecomment-702052403 ?
I don't exactly understand what to do in the suricata config https://github.com/StamusNetworks/SELKS/issues/255#issuecomment-698755156
Maybe here is the way to solve the problem.
I updated - but can't seem to reproduce it. Which old config files do you exactly keep?
I keeped all old config files.
Dne 05. 10. 20 v 8:40 Peter Manev napsal(a):
I updated - but cant seem to reproduce it. Which old config files do you exactly keep?
— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/StamusNetworks/SELKS/issues/251#issuecomment-703431517, or unsubscribe https://github.com/notifications/unsubscribe-auth/ABGUNEJJ6QJ2MLU4QIBE5YDSJFS67ANCNFSM4QLWP6DA.
-- Ing. Michal Vymazal Senior Cyber Security Architect Linux Services CEO
vymazal@linuxservices.cz www.linuxservices.cz
Home Computer
Druzstvo kyberneticke obrany Predseda predstavenstva druzstvo.kybernetickaobrana.cz
LinkedIn profile https://www.linkedin.com/in/linuxservices/
This mail can't contain any virus. I'm using only Open Source software.
Today I updated one selks device and the result is exact the same as
https://github.com/StamusNetworks/SELKS/issues/251#issuecomment-702720837
I updated all config files too.
What is the current status - only Scirius management pages do not populate or ? Does Hunt populate?
Hunt populate is working. Here is the problem with Discover - Elastic
I also copied the response to the clipboard, attached in the file clipboard.txt clipboard.txt
ok, having a look and proposing a way forward.
Might be unrelated but I also have a blank white page for Kibana. It's a brand new install from ISO. What seems to be happening is a 404 error in the web console where it can't find bootstrap.js.
Did you go over the first time setup wiki - I am thinking might be related to - https://github.com/StamusNetworks/SELKS/wiki/First-time-setup#nginx-config
Did you go over the first time setup wiki - I am thinking might be related to - https://github.com/StamusNetworks/SELKS/wiki/First-time-setup#nginx-config
That took care of it for me. I missed that, simply assuming that it was ok as I went through everything but clicking that link. 👍 Thanks.
Glad to hear it was fixed! This will be addressed in the next release upgrade too. Thank you
The script selks-upgrade_stamus
passed OK. But after running the script selks-first-time-setup_stamus