graphaware / neo4j-to-elasticsearch

GraphAware Framework Module for Integrating Neo4j with Elasticsearch
261 stars 57 forks source link

Unable to index ERROR Could not deserialize metadata for module ID ES #151

Closed itaybraha closed 3 years ago

itaybraha commented 5 years ago

Hi,

I am experiencing a strange issue when I try to start a docker causal cluster (see docker-compose.yml as well as conf files below), after having loaded a database using neo4j-admin and trying to get it re-indexd in Elasticsearch. What is strange to me is that the same configuration works when on one node on my machine and fails with the following message on a docker cluster:

Unable to index ERROR Could not deserialize metadata for module ID ES

Needless to say there are no errors with the same configuration running on a single node on my machine.

I am using Neo4j 3.5.1 enterprise version to be fair and am not sure if it is yet supported.

Note: the external lan docker network was pre-created to serve both Neo4j and Elasticsearch docker cluster in communicating with each other.

docker-compose.yml


networks:
  default:
    external:
      name: lan
services:
  core1:
    image: neo4j:3.5.1-enterprise
    container_name: core-placeholder
    stdin_open: true
    tty: true
    networks:
      - default
    ports:
      - "7475:7475"
      - "7688:7688" 
      - "6477:6477"
    volumes:
      - ./conf01:/var/lib/neo4j/conf
      - ./plugins01:/var/lib/neo4j/plugins
      - ./data01:/var/lib/neo4j/data
      - ./import:/var/lib/neo4j/import
      - ./run:/var/lib/neo4j/run
    environment:
      - NEO4J_ACCEPT_LICENSE_AGREEMENT=yes
      - NEO4J_dbms_connectors_default__listen__address=0.0.0.0
      - NEO4J_dbms_memory_pagecache_size=512M
      - NEO4J_dbms_memory_heap_initial__size=512M
      - NEO4J_dbms_mode=CORE
      - NEO4J_causal__clustering_minimum__core__cluster__size__at__formation=3
      - NEO4J_causalClustering_initialDiscoveryMembers=core1:5000,core2:5000,core3:5000
      - NEO4J_dbms_connector_http_listen__address=:7475
      - NEO4J_dbms_connector_https_listen__address=:6477
      - NEO4J_dbms_connector_bolt_listen__address=:7688
      - NEO4J_dbms_logs_security_level=DEBUG

  core2:
    image: neo4j:3.5.1-enterprise
    stdin_open: true
    tty: true
    networks:
      - default
    ports:
      - "7476:7476"
      - "7689:7689" 
      - "6478:6478"
    volumes:
      - ./conf02:/var/lib/neo4j/conf
      - ./plugins02:/var/lib/neo4j/plugins
      - ./data02:/var/lib/neo4j/data
      - ./import:/var/lib/neo4j/import
      - ./run:/var/lib/neo4j/run
    environment:
      - NEO4J_ACCEPT_LICENSE_AGREEMENT=yes
      - NEO4J_dbms_connectors_default__listen__address=0.0.0.0
      - NEO4J_dbms_memory_pagecache_size=512M
      - NEO4J_dbms_memory_heap_initial__size=512M
      - NEO4J_dbms_mode=CORE
      - NEO4J_causal__clustering_minimum__core__cluster__size__at__formation=3
      - NEO4J_causalClustering_initialDiscoveryMembers=core1:5000,core2:5000,core3:5000
      - NEO4J_dbms_connector_http_listen__address=:7476
      - NEO4J_dbms_connector_https_listen__address=:6478
      - NEO4J_dbms_connector_bolt_listen__address=:7689
      - NEO4J_dbms_logs_security_level=DEBUG

  core3:
    image: neo4j:3.5.1-enterprise
    stdin_open: true
    tty: true
    networks:
      - default
    ports:
      - "7477:7477"
      - "7690:7690" 
      - "6479:6479"
    volumes:
      - ./conf03:/var/lib/neo4j/conf
      - ./plugins03:/var/lib/neo4j/plugins
      - ./data03:/var/lib/neo4j/data
      - ./import:/var/lib/neo4j/import
      - ./run:/var/lib/neo4j/run
    environment:
      - NEO4J_ACCEPT_LICENSE_AGREEMENT=yes
      - NEO4J_dbms_connectors_default__listen__address=0.0.0.0
      - NEO4J_dbms_memory_pagecache_size=512M
      - NEO4J_dbms_memory_heap_initial__size=512M
      - NEO4J_dbms_mode=CORE
      - NEO4J_causalClustering_discoveryAdvertisedAddress=core3:5000
      - NEO4J_causalClustering_transactionAdvertisedAddress=core3:6000
      - NEO4J_causalClustering_raftAdvertisedAddress=core3:7000
      - NEO4J_causal_clustering__minimum__core__cluster__size__at__formation=3
      - NEO4J_causalClustering_initialDiscoveryMembers=core1:5000,core2:5000,core3:5000
      - NEO4J_dbms_connector_http_listen__address=:7477
      - NEO4J_dbms_connector_https_listen__address=:6479      
      - NEO4J_dbms_connector_bolt_listen__address=:7690
      - NEO4J_dbms_logs_security_level=DEBUG

neo4j.conf:

#********************************************************************
# ES Neo4j
#********************************************************************

# This setting should only be set once for registering the framework and all the used submodules
dbms.unmanaged_extension_classes=com.graphaware.server=/graphaware

com.graphaware.runtime.enabled=true

#UIDM becomes the module ID:
com.graphaware.module.UIDM.1=com.graphaware.module.uuid.UuidBootstrapper

#optional, default is "uuid". (only if using the UUID module)
com.graphaware.module.UIDM.uuidProperty=_myuniqueidname

#optional, default is all nodes:
#com.graphaware.module.UIDM.node=hasLabel('Label1') || hasLabel('Label2')

#optional, default is uuidIndex
com.graphaware.module.UIDM.uuidIndex=uuidIndex

#prevent the whole db to be assigned a new uuid if the uuid module is settle up together with neo4j2es
com.graphaware.module.UIDM.initializeUntil=1556215200000 (in the future)

#ES becomes the module ID:
com.graphaware.module.ES.2=com.graphaware.module.es.ElasticSearchModuleBootstrapper

#URI of Elasticsearch
com.graphaware.module.ES.uri=elastic_ip

#Port of Elasticsearch
com.graphaware.module.ES.port=9200

#optional, protocol of Elasticsearch connection, defaults to http
com.graphaware.module.ES.protocol=http

#optional, Elasticsearch index name, default is neo4j-index
#com.graphaware.module.ES.index=my_data_index

#optional, node property key of a propery that is used as unique identifier of the node. Must be the same as com.graphaware.module.UIDM.uuidProperty (only if using UUID module), defaults to uuid
#use "ID()" to use native Neo4j IDs as Elasticsearch IDs (not recommended)
#com.graphaware.module.ES.keyProperty=_myuniqueidname

#optional, whether to retry if a replication fails, defaults to false
com.graphaware.module.ES.retryOnError=true

#optional, size of the in-memory queue that queues up operations to be synchronised to Elasticsearch, defaults to 10000
com.graphaware.module.ES.queueSize=10000

#optional, size of the batch size to use during re-initialization, defaults to 1000
com.graphaware.module.ES.reindexBatchSize=200

#optional, specify which nodes to index in Elasticsearch, defaults to all nodes
#com.graphaware.module.ES.node=hasLabel('context')

#optional, specify which node properties to index in Elasticsearch, defaults to all properties
#com.graphaware.module.ES.node.property=key != 'age'

#optional, specify whether to send updates to Elasticsearch in bulk, defaults to true (highly recommended)
com.graphaware.module.ES.bulk=false

#optional, read explanation below, defaults to 0
com.graphaware.module.ES.initializeUntil=1556215200000 (in the future)

#optional, whether or not the reindexation process (when db start) should be made in asynchronous mode
#default is "false" and the db will not be available until completed
#com.graphaware.module.ES.asyncIndexation=true

com.graphaware.module.ES.keyProperty=_myuniqueidname
com.graphaware.module.ES.relationship=(true)
com.graphaware.module.ES.mapping=com.graphaware.module.es.mapping.JsonFileMapping
com.graphaware.module.ES.file=mapping.json

#####################################################################
# UUID EXTRA
#####################################################################

#optional, default is no relationships:
#com.graphaware.module.UIDM.relationship=isType('Type1')

#optional, default is uuidRelIndex
com.graphaware.module.UIDM.uuidRelationshipIndex=uuidRelIndex

#assign uuid's to all relationships
com.graphaware.module.UIDM.relationship=com.graphaware.runtime.policy.all.IncludeAllBusinessRelationships

mapping.json

{
    "defaults": {
        "key_property": "_myuniqueidname",
        "nodes_index": "node_idx",
        "relationships_index": "rel_idx",
        "include_remaining_properties": true,
        "blacklisted_node_properties": ["prop1","prop2","prop4","prop5"]

    },
    "node_mappings": [{
            "condition": "hasLabel('label1') || hasLabel('label2') || hasLabel('label3') || hasLabel('label4') || hasLabel('label5')",
            "type": "nodes",
            "properties": {
                "tags": "getLabels()"
            }
        }
    ],
    "relationship_mappings": [{
                "condition": "isType('rel1') || isType('rel2') || isType('rel3') || isType('rel4') || isType('rel5') || isType('rel6') || isType('rel7') || isType('rel8')",
                "type": "reln",
            "properties":{
                        "relation": "getType()"
                 }}
    ]
}

Any help will be appreciated..

ikwattro commented 5 years ago

Does it throws the exception on all the cores ?

itaybraha commented 5 years ago

Yes.

itaybraha commented 5 years ago

@ikwattro any idea what that may indicate?

ikwattro commented 5 years ago

Didn't have time to look at it yet, can do this weekend.

On Fri, 26 Apr 2019, 23:13 Itay Braha, notifications@github.com wrote:

@ikwattro https://github.com/ikwattro any idea what that may indicate?

— You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/graphaware/neo4j-to-elasticsearch/issues/151#issuecomment-487203213, or mute the thread https://github.com/notifications/unsubscribe-auth/AAJKK6P7JKAHSRXRKJGTBTTPSNV7BANCNFSM4HHY774A .

ikwattro commented 5 years ago

So, I could not reproduce. Any chance you can send us your stores as well as config so we can investigate ?

itaybraha commented 5 years ago

@ikwattro sending the stores is a bit of an issue at the moment. I have no issue sharing the config and full logs.

neo4j.conf

#*****************************************************************
# Neo4j configuration
#
# For more details and a complete list of settings, please see
# https://neo4j.com/docs/operations-manual/current/reference/configuration-settings/
#*****************************************************************

# The name of the database to mount. Note that this is *not* to be confused with
# the causal_clustering.database setting, used to specify a logical database
# name when creating a multi-clustering deployment.
#dbms.active_database=graph.db

# Paths of directories in the installation.
#dbms.directories.data=/var/lib/neo4j/data
#dbms.directories.plugins=/var/lib/neo4j/plugins
#dbms.directories.certificates=/var/lib/neo4j/certificates
#dbms.directories.logs=/var/log/neo4j
#dbms.directories.lib=/usr/share/neo4j/lib
#dbms.directories.run=/var/run/neo4j
#dbms.directories.metrics=/var/lib/neo4j/metrics

# This setting constrains all `LOAD CSV` import files to be under the `import` directory. Remove or comment it out to
# allow files to be loaded from anywhere in the filesystem; this introduces possible security problems. See the
# `LOAD CSV` section of the manual for details.
dbms.directories.import=/var/lib/neo4j/import

# Whether requests to Neo4j are authenticated.
# To disable authentication, uncomment this line
dbms.security.auth_enabled=false

# Enable this to be able to upgrade a store from an older version.
dbms.allow_upgrade=true

# Java Heap Size: by default the Java heap size is dynamically
# calculated based on available system resources.
# Uncomment these lines to set specific initial and maximum
# heap size.
#dbms.memory.heap.initial_size=512m
#dbms.memory.heap.max_size=512m

# The amount of memory to use for mapping the store files, in bytes (or
# kilobytes with the 'k' suffix, megabytes with 'm' and gigabytes with 'g').
# If Neo4j is running on a dedicated server, then it is generally recommended
# to leave about 2-4 gigabytes for the operating system, give the JVM enough
# heap to hold all your transaction state and query context, and then leave the
# rest for the page cache.
# The default page cache memory assumes the machine is dedicated to running
# Neo4j, and is heuristically set to 50% of RAM minus the max Java heap size.
#dbms.memory.pagecache.size=10g

# Enable online backups to be taken from this database.
#dbms.backup.enabled=true

# By default the backup service will only listen on localhost.
# To enable remote backups you will have to bind to an external
# network interface (e.g. 0.0.0.0 for all interfaces).
# The protocol running varies depending on deployment. In a Causal Clustering environment this is the
# same protocol that runs on causal_clustering.transaction_listen_address.
#dbms.backup.address={{ node_ip_address }}:6362

# Enable encryption on the backup service for CC instances (does not work for single-instance or HA clusters)
#dbms.backup.ssl_policy=backup

#*****************************************************************
# Network connector configuration
#*****************************************************************

# With default configuration Neo4j only accepts local connections.
# To accept non-local connections, uncomment this line:
#dbms.connectors.default_listen_address=0.0.0.0

# You can also choose a specific network interface, and configure a non-default
# port for each connector, by setting their individual listen_address.

# The address at which this server can be reached by its clients. This may be the server's IP address or DNS name, or
# it may be the address of a reverse proxy which sits in front of the server. This setting may be overridden for
# individual connectors below.
#dbms.connectors.default_advertised_address={{ node_ip_address }}

# You can also choose a specific advertised hostname or IP address, and
# configure an advertised port for each connector, by setting their
# individual advertised_address.

# Bolt connector
dbms.connector.bolt.enabled=true
dbms.connector.bolt.tls_level=OPTIONAL
#dbms.connector.bolt.listen_address=:7687

# HTTP Connector. There must be exactly one HTTP connector.
dbms.connector.http.enabled=true
#dbms.connector.http.listen_address=:7474

# HTTPS Connector. There can be zero or one HTTPS connectors.
dbms.connector.https.enabled=true
#dbms.connector.https.listen_address=:7473

# Number of Neo4j worker threads.
#dbms.threads.worker_count=

#*****************************************************************
# SSL system configuration
#*****************************************************************

# Names of the SSL policies to be used for the respective components.

# The legacy policy is a special policy which is not defined in
# the policy configuration section, but rather derives from
# dbms.directories.certificates and associated files
# (by default: neo4j.key and neo4j.cert). Its use will be deprecated.

# The policies to be used for connectors.
#
# N.B: Note that a connector must be configured to support/require
#      SSL/TLS for the policy to actually be utilized.
#
# see: dbms.connector.*.tls_level

#bolt.ssl_policy=legacy
#https.ssl_policy=legacy

# For a causal cluster the configuring of a policy mandates its use.

#causal_clustering.ssl_policy=

#*****************************************************************
# SSL policy configuration
#*****************************************************************

# Each policy is configured under a separate namespace, e.g.
#    dbms.ssl.policy.<policyname>.*
#
# The example settings below are for a new policy named 'default'.

# The base directory for cryptographic objects. Each policy will by
# default look for its associated objects (keys, certificates, ...)
# under the base directory.
#
# Every such setting can be overriden using a full path to
# the respective object, but every policy will by default look
# for cryptographic objects in its base location.
#
# Mandatory setting

#dbms.ssl.policy.default.base_directory=certificates/default

# Allows the generation of a fresh private key and a self-signed
# certificate if none are found in the expected locations. It is
# recommended to turn this off again after keys have been generated.
#
# Keys should in general be generated and distributed offline
# by a trusted certificate authority (CA) and not by utilizing
# this mode.

#dbms.ssl.policy.default.allow_key_generation=false

# Enabling this makes it so that this policy ignores the contents
# of the trusted_dir and simply resorts to trusting everything.
#
# Use of this mode is discouraged. It would offer encryption but no security.

#dbms.ssl.policy.default.trust_all=false

# The private key for the default SSL policy. By default a file
# named private.key is expected under the base directory of the policy.
# It is mandatory that a key can be found or generated.

#dbms.ssl.policy.default.private_key=

# The private key for the default SSL policy. By default a file
# named public.crt is expected under the base directory of the policy.
# It is mandatory that a certificate can be found or generated.

#dbms.ssl.policy.default.public_certificate=

# The certificates of trusted parties. By default a directory named
# 'trusted' is expected under the base directory of the policy. It is
# mandatory to create the directory so that it exists, because it cannot
# be auto-created (for security purposes).
#
# To enforce client authentication client_auth must be set to 'require'!

#dbms.ssl.policy.default.trusted_dir=

# Certificate Revocation Lists (CRLs). By default a directory named
# 'revoked' is expected under the base directory of the policy. It is
# mandatory to create the directory so that it exists, because it cannot
# be auto-created (for security purposes).

#dbms.ssl.policy.default.revoked_dir=

# Client authentication setting. Values: none, optional, require
# The default is to require client authentication.
#
# Servers are always authenticated unless explicitly overridden
# using the trust_all setting. In a mutual authentication setup this
# should be kept at the default of require and trusted certificates
# must be installed in the trusted_dir.

#dbms.ssl.policy.default.client_auth=require

# A comma-separated list of allowed TLS versions.
# By default only TLSv1.2 is allowed.

#dbms.ssl.policy.default.tls_versions=

# A comma-separated list of allowed ciphers.
# The default ciphers are the defaults of the JVM platform.

#dbms.ssl.policy.default.ciphers=

#*****************************************************************
# Logging configuration
#*****************************************************************

# To enable HTTP logging, uncomment this line
dbms.logs.http.enabled=true

# Number of HTTP logs to keep.
#dbms.logs.http.rotation.keep_number=5

# Size of each HTTP log that is kept.
#dbms.logs.http.rotation.size=20m

# To enable GC Logging, uncomment this line
#dbms.logs.gc.enabled=true

# GC Logging Options
# see http://docs.oracle.com/cd/E19957-01/819-0084-10/pt_tuningjava.html#wp57013 for more information.
#dbms.logs.gc.options=-XX:+PrintGCDetails -XX:+PrintGCDateStamps -XX:+PrintGCApplicationStoppedTime -XX:+PrintPromotionFailure -XX:+PrintTenuringDistribution

# Number of GC logs to keep.
#dbms.logs.gc.rotation.keep_number=5

# Size of each GC log that is kept.
#dbms.logs.gc.rotation.size=20m

# Size threshold for rotation of the debug log. If set to zero then no rotation will occur. Accepts a binary suffix "k",
# "m" or "g".
#dbms.logs.debug.rotation.size=20m

# Maximum number of history files for the internal log.
#dbms.logs.debug.rotation.keep_number=7

# Log executed queries that takes longer than the configured threshold. Enable by uncommenting this line.
#dbms.logs.query.enabled=true

# If the execution of query takes more time than this threshold, the query is logged. If set to zero then all queries
# are logged.
#dbms.logs.query.threshold=0

# The file size in bytes at which the query log will auto-rotate. If set to zero then no rotation will occur. Accepts a
# binary suffix "k", "m" or "g".
#dbms.logs.query.rotation.size=20m

# Maximum number of history files for the query log.
#dbms.logs.query.rotation.keep_number=7

# Include parameters for the executed queries being logged (this is enabled by default).
#dbms.logs.query.parameter_logging_enabled=true

# Uncomment this line to include detailed time information for the executed queries being logged:
#dbms.logs.query.time_logging_enabled=true

# Uncomment this line to include bytes allocated by the executed queries being logged:
#dbms.logs.query.allocation_logging_enabled=true

# Uncomment this line to include page hits and page faults information for the executed queries being logged:
#dbms.logs.query.page_logging_enabled=true

# The security log is always enabled when `dbms.security.auth_enabled=true`, and resides in `logs/security.log`.

# Log level for the security log. One of DEBUG, INFO, WARN and ERROR.
dbms.logs.security.level=DEBUG

# Threshold for rotation of the security log.
#dbms.logs.security.rotation.size=20m

# Minimum time interval after last rotation of the security log before it may be rotated again.
#dbms.logs.security.rotation.delay=300s

# Maximum number of history files for the security log.
#dbms.logs.security.rotation.keep_number=7

#*****************************************************************
# Causal Clustering Configuration
#*****************************************************************

# Uncomment and specify these lines for running Neo4j in Causal Clustering mode.
# See the Causal Clustering documentation at https://neo4j.com/docs/ for details.

# Database mode
# Allowed values:
# CORE - Core member of the cluster, part of the consensus quorum.
# READ_REPLICA - Read replica in the cluster, an eventually-consistent read-only instance of the database.
# To operate this Neo4j instance in Causal Clustering mode as a core member, uncomment this line:
#dbms.mode=CORE

# Expected number of Core servers in the cluster at formation
#causal_clustering.minimum_core_cluster_size_at_formation=3

# Minimum expected number of Core servers in the cluster at runtime.
#causal_clustering.minimum_core_cluster_size_at_runtime=3

# Savas
#causal_clustering.expected_core_cluster_size=3

# A comma-separated list of the address and port for which to reach all other members of the cluster. It must be in the
# host:port format. For each machine in the cluster, the address will usually be the public ip address of that machine.
# The port will be the value used in the setting "causal_clustering.discovery_listen_address".
#causal_clustering.initial_discovery_members=core1:5000,core2:5000,core3:5000

# Host and port to bind the cluster member discovery management communication.
# This is the setting to add to the collection of address in causal_clustering.initial_core_cluster_members.
# Use 0.0.0.0 to bind to any network interface on the machine. If you want to only use a specific interface
# (such as a private ip address on AWS, for example) then use that ip address instead.
# If you don't know what value to use here, use this machines ip address.
#causal_clustering.discovery_listen_address=:5000

# Network interface and port for the transaction shipping server to listen on.
# Please note that it is also possible to run the backup client against this port so always limit access to it via the
# firewall and configure an ssl policy. If you want to allow for messages to be read from
# any network on this machine, us 0.0.0.0. If you want to constrain communication to a specific network address
# (such as a private ip on AWS, for example) then use that ip address instead.
# If you don't know what value to use here, use this machines ip address.
#causal_clustering.transaction_listen_address=:6000

# Network interface and port for the RAFT server to listen on. If you want to allow for messages to be read from
# any network on this machine, us 0.0.0.0. If you want to constrain communication to a specific network address
# (such as a private ip on AWS, for example) then use that ip address instead.
# If you don't know what value to use here, use this machines ip address.
#causal_clustering.raft_listen_address=:7000

# List a set of names for groups to which this server should belong. This
# is a comma-separated list and names should only use alphanumericals
# and underscore. This can be used to identify groups of servers in the
# configuration for load balancing and replication policies.
#
# The main intention for this is to group servers, but it is possible to specify
# a unique identifier here as well which might be useful for troubleshooting
# or other special purposes.
#causal_clustering.server_groups=

#*****************************************************************
# Causal Clustering Load Balancing
#*****************************************************************

# N.B: Read the online documentation for a thorough explanation!

# Selects the load balancing plugin that shall be enabled.
#causal_clustering.load_balancing.plugin=server_policies

####### Examples for "server_policies" plugin #######

# Will select all available servers as the default policy, which is the
# policy used when the client does not specify a policy preference. The
# default configuration for the default policy is all().
#causal_clustering.load_balancing.config.server_policies.default=all()

# Will select servers in groups 'group1' or 'group2' under the default policy.
#causal_clustering.load_balancing.config.server_policies.default=groups(group1,group2)

# Slightly more advanced example:
# Will select servers in 'group1', 'group2' or 'group3', but only if there are at least 2.
# This policy will be exposed under the name of 'mypolicy'.
#causal_clustering.load_balancing.config.server_policies.mypolicy=groups(group1,group2,group3) -> min(2)

# Below will create an even more advanced policy named 'regionA' consisting of several rules
# yielding the following behaviour:
#
#            select servers in regionA, if at least 2 are available
# otherwise: select servers in regionA and regionB, if at least 2 are available
# otherwise: select all servers
#
# The intention is to create a policy for a particular region which prefers
# a certain set of local servers, but which will fallback to other regions
# or all available servers as required.
#
# N.B: The following configuration uses the line-continuation character \
#      which allows you to construct an easily readable rule set spanning
#      several lines.
#
#causal_clustering.load_balancing.config.server_policies.policyA=\
#groups(regionA) -> min(2);\
#groups(regionA,regionB) -> min(2);

# Note that implicitly the last fallback is to always consider all() servers,
# but this can be prevented by specifying a halt() as the last rule.
#
#causal_clustering.load_balancing.config.server_policies.regionA_only=\
#groups(regionA);\
#halt();

#*****************************************************************
# Causal Clustering Additional Configuration Options
#*****************************************************************
# The following settings are used less frequently.
# If you don't know what these are, you don't need to change these from their default values.

# The name of the database being hosted by this server instance. This 
# configuration setting may be safely ignored unless deploying a multicluster. 
# Instances may be allocated to constituent clusters by assigning them 
# distinct database names using this setting. For instance if you had 6 
# instances you could form 2 clusters by assigning half the database name 
# "foo", half the name "bar". The setting value must match exactly between 
# members of the same cluster. This setting is a one-off: once an instance 
# is configured with a database name it may not be changed in future without
# using `neo4j-admin unbind`. 
#causal_clustering.database=default 

# Address and port that this machine advertises that it's RAFT server is listening at. Should be a
# specific network address. If you are unsure about what value to use here, use this machine's ip address.
#causal_clustering.raft_advertised_address=:7000

# Address and port that this machine advertises that it's transaction shipping server is listening at. Should be a
# specific network address. If you are unsure about what value to use here, use this machine's ip address.
#causal_clustering.transaction_advertised_address=:6000

# The time limit within which a new leader election will occur if no messages from the current leader are received.
# Larger values allow for more stable leaders at the expense of longer unavailability times in case of leader
# failures.
#causal_clustering.leader_election_timeout=7s

# The time limit allowed for a new member to attempt to update its data to match the rest of the cluster.
#causal_clustering.join_catch_up_timeout=10m

# The size of the batch for streaming entries to other machines while trying to catch up another machine.
#causal_clustering.catchup_batch_size=64

# When to pause sending entries to other machines and allow them to catch up.
#causal_clustering.log_shipping_max_lag=256

# Raft log pruning frequncy.
#causal_clustering.raft_log_pruning_frequency=10m

# The size to allow the raft log to grow before rotating.
#causal_clustering.raft_log_rotation_size=250M

### The following setting is relevant for Edge servers only.
# The interval of pulling updates from Core servers.
#causal_clustering.pull_interval=1s

# For how long should drivers cache the discovery data from
# the dbms.cluster.routing.getServers() procedure. Defaults to 300s.
#causal_clustering.cluster_routing_ttl=300s

#*****************************************************************
# HA configuration
#*****************************************************************

# Uncomment and specify these lines for running Neo4j in High Availability mode.
# See the High Availability documentation at https://neo4j.com/docs/ for details.

# Database mode
# Allowed values:
# HA - High Availability
# SINGLE - Single mode, default.
# To run in High Availability mode uncomment this line:
#dbms.mode=HA

# ha.server_id is the number of each instance in the HA cluster. It should be
# an integer (e.g. 1), and should be unique for each cluster instance.
#ha.server_id={{ neo4j_server_id }}

# ha.initial_hosts is a comma-separated list (without spaces) of the host:port
# where the ha.host.coordination of all instances will be listening. Typically
# this will be the same for all cluster instances.
#ha.initial_hosts={{ neo4j_initial_hosts }}

# IP and port for this instance to listen on, for communicating cluster status
# information with other instances (also see ha.initial_hosts). The IP
# must be the configured IP address for one of the local interfaces.
#ha.host.coordination={{ node_ip_address }}:5001

# IP and port for this instance to listen on, for communicating transaction
# data with other instances (also see ha.initial_hosts). The IP
# must be the configured IP address for one of the local interfaces.
#ha.host.data={{ node_ip_address }}:6001

# The interval, in seconds, at which slaves will pull updates from the master. You must comment out
# the option to disable periodic pulling of updates.
#ha.pull_interval=10

# Amount of slaves the master will try to push a transaction to upon commit
# (default is 1). The master will optimistically continue and not fail the
# transaction even if it fails to reach the push factor. Setting this to 0 will
# increase write performance when writing through master but could potentially
# lead to branched data (or loss of transaction) if the master goes down.
#ha.tx_push_factor=1

# Strategy the master will use when pushing data to slaves (if the push factor
# is greater than 0). There are three options available "fixed_ascending" (default),
# "fixed_descending" or "round_robin". Fixed strategies will start by pushing to
# slaves ordered by server id (accordingly with qualifier) and are useful when
# planning for a stable fail-over based on ids.
#ha.tx_push_strategy=fixed_ascending

# Policy for how to handle branched data.
#ha.branched_data_policy=keep_all

# How often heartbeat messages should be sent. Defaults to ha.default_timeout.
#ha.heartbeat_interval=5s

# How long to wait for heartbeats from other instances before marking them as suspects for failure.
# This value reflects considerations of network latency, expected duration of garbage collection pauses
# and other factors that can delay message sending and processing. Larger values will result in more
# stable masters but also will result in longer waits before a failover in case of master failure.
# This value should not be set to less than twice the ha.heartbeat_interval value otherwise there is a high
# risk of frequent master switches and possibly branched data occurrence.
#ha.heartbeat_timeout=40s

# If you are using a load-balancer that doesn't support HTTP Auth, you may need to turn off authentication for the
# HA HTTP status endpoint by uncommenting the following line.
#dbms.security.ha_status_auth_enabled=false

# Whether this instance should only participate as slave in cluster. If set to
# true, it will never be elected as master.
#ha.slave_only=false

#********************************************************************
# Security Configuration
#********************************************************************

# The authentication and authorization provider that contains both users and roles.
# This can be one of the built-in `native` or `ldap` auth providers,
# or it can be an externally provided plugin, with a custom name prefixed by `plugin`,
# i.e. `plugin-<AUTH_PROVIDER_NAME>`.
#dbms.security.auth_provider=native

# The time to live (TTL) for cached authentication and authorization info when using
# external auth providers (LDAP or plugin). Setting the TTL to 0 will
# disable auth caching.
#dbms.security.auth_cache_ttl=10m

# The maximum capacity for authentication and authorization caches (respectively).
#dbms.security.auth_cache_max_capacity=10000

# Set to log successful authentication events to the security log.
# If this is set to `false` only failed authentication events will be logged, which
# could be useful if you find that the successful events spam the logs too much,
# and you do not require full auditing capability.
#dbms.security.log_successful_authentication=true

#================================================
# LDAP Auth Provider Configuration
#================================================

# URL of LDAP server to use for authentication and authorization.
# The format of the setting is `<protocol>://<hostname>:<port>`, where hostname is the only required field.
# The supported values for protocol are `ldap` (default) and `ldaps`.
# The default port for `ldap` is 389 and for `ldaps` 636.
# For example: `ldaps://ldap.example.com:10389`.
#
# NOTE: You may want to consider using STARTTLS (`dbms.security.ldap.use_starttls`) instead of LDAPS
# for secure connections, in which case the correct protocol is `ldap`.
#dbms.security.ldap.host=localhost

# Use secure communication with the LDAP server using opportunistic TLS.
# First an initial insecure connection will be made with the LDAP server, and then a STARTTLS command
# will be issued to negotiate an upgrade of the connection to TLS before initiating authentication.
#dbms.security.ldap.use_starttls=false

# The LDAP referral behavior when creating a connection. This is one of `follow`, `ignore` or `throw`.
# `follow` automatically follows any referrals
# `ignore` ignores any referrals
# `throw` throws an exception, which will lead to authentication failure
#dbms.security.ldap.referral=follow

# The timeout for establishing an LDAP connection. If a connection with the LDAP server cannot be
# established within the given time the attempt is aborted.
# A value of 0 means to use the network protocol's (i.e., TCP's) timeout value.
#dbms.security.ldap.connection_timeout=30s

# The timeout for an LDAP read request (i.e. search). If the LDAP server does not respond within
# the given time the request will be aborted. A value of 0 means wait for a response indefinitely.
#dbms.security.ldap.read_timeout=30s

#----------------------------------
# LDAP Authentication Configuration
#----------------------------------

# LDAP authentication mechanism. This is one of `simple` or a SASL mechanism supported by JNDI,
# for example `DIGEST-MD5`. `simple` is basic username
# and password authentication and SASL is used for more advanced mechanisms. See RFC 2251 LDAPv3
# documentation for more details.
#dbms.security.ldap.authentication.mechanism=simple

# LDAP user DN template. An LDAP object is referenced by its distinguished name (DN), and a user DN is
# an LDAP fully-qualified unique user identifier. This setting is used to generate an LDAP DN that
# conforms with the LDAP directory's schema from the user principal that is submitted with the
# authentication token when logging in.
# The special token {0} is a placeholder where the user principal will be substituted into the DN string.
#dbms.security.ldap.authentication.user_dn_template=uid={0},ou=users,dc=example,dc=com

# Determines if the result of authentication via the LDAP server should be cached or not.
# Caching is used to limit the number of LDAP requests that have to be made over the network
# for users that have already been authenticated successfully. A user can be authenticated against
# an existing cache entry (instead of via an LDAP server) as long as it is alive
# (see `dbms.security.auth_cache_ttl`).
# An important consequence of setting this to `true` is that
# Neo4j then needs to cache a hashed version of the credentials in order to perform credentials
# matching. This hashing is done using a cryptographic hash function together with a random salt.
# Preferably a conscious decision should be made if this method is considered acceptable by
# the security standards of the organization in which this Neo4j instance is deployed.
#dbms.security.ldap.authentication.cache_enabled=true

#----------------------------------
# LDAP Authorization Configuration
#----------------------------------
# Authorization is performed by searching the directory for the groups that
# the user is a member of, and then map those groups to Neo4j roles.

# Perform LDAP search for authorization info using a system account instead of the user's own account.
#
# If this is set to `false` (default), the search for group membership will be performed
# directly after authentication using the LDAP context bound with the user's own account.
# The mapped roles will be cached for the duration of `dbms.security.auth_cache_ttl`,
# and then expire, requiring re-authentication. To avoid frequently having to re-authenticate
# sessions you may want to set a relatively long auth cache expiration time together with this option.
# NOTE: This option will only work if the users are permitted to search for their
# own group membership attributes in the directory.
#
# If this is set to `true`, the search will be performed using a special system account user
# with read access to all the users in the directory.
# You need to specify the username and password using the settings
# `dbms.security.ldap.authorization.system_username` and
# `dbms.security.ldap.authorization.system_password` with this option.
# Note that this account only needs read access to the relevant parts of the LDAP directory
# and does not need to have access rights to Neo4j, or any other systems.
#dbms.security.ldap.authorization.use_system_account=false

# An LDAP system account username to use for authorization searches when
# `dbms.security.ldap.authorization.use_system_account` is `true`.
# Note that the `dbms.security.ldap.authentication.user_dn_template` will not be applied to this username,
# so you may have to specify a full DN.
#dbms.security.ldap.authorization.system_username=

# An LDAP system account password to use for authorization searches when
# `dbms.security.ldap.authorization.use_system_account` is `true`.
#dbms.security.ldap.authorization.system_password=

# The name of the base object or named context to search for user objects when LDAP authorization is enabled.
# A common case is that this matches the last part of `dbms.security.ldap.authentication.user_dn_template`.
#dbms.security.ldap.authorization.user_search_base=ou=users,dc=example,dc=com

# The LDAP search filter to search for a user principal when LDAP authorization is
# enabled. The filter should contain the placeholder token {0} which will be substituted for the
# user principal.
#dbms.security.ldap.authorization.user_search_filter=(&(objectClass=*)(uid={0}))

# A list of attribute names on a user object that contains groups to be used for mapping to roles
# when LDAP authorization is enabled.
#dbms.security.ldap.authorization.group_membership_attributes=memberOf

# An authorization mapping from LDAP group names to Neo4j role names.
# The map should be formatted as a semicolon separated list of key-value pairs, where the
# key is the LDAP group name and the value is a comma separated list of corresponding role names.
# For example: group1=role1;group2=role2;group3=role3,role4,role5
#
# You could also use whitespaces and quotes around group names to make this mapping more readable,
# for example: dbms.security.ldap.authorization.group_to_role_mapping=\
#          "cn=Neo4j Read Only,cn=users,dc=example,dc=com"      = reader;    \
#          "cn=Neo4j Read-Write,cn=users,dc=example,dc=com"     = publisher; \
#          "cn=Neo4j Schema Manager,cn=users,dc=example,dc=com" = architect; \
#          "cn=Neo4j Administrator,cn=users,dc=example,dc=com"  = admin
#dbms.security.ldap.authorization.group_to_role_mapping=

#*****************************************************************
# Miscellaneous configuration
#*****************************************************************

# Enable this to specify a parser other than the default one.
#cypher.default_language_version=3.0

# Determines if Cypher will allow using file URLs when loading data using
# `LOAD CSV`. Setting this value to `false` will cause Neo4j to fail `LOAD CSV`
# clauses that load data from the file system.
#dbms.security.allow_csv_import_from_file_urls=true

# Retention policy for transaction logs needed to perform recovery and backups.
#dbms.tx_log.rotation.retention_policy=7 days

# Limit the number of IOs the background checkpoint process will consume per second.
# This setting is advisory, is ignored in Neo4j Community Edition, and is followed to
# best effort in Enterprise Edition.
# An IO is in this case a 8 KiB (mostly sequential) write. Limiting the write IO in
# this way will leave more bandwidth in the IO subsystem to service random-read IOs,
# which is important for the response time of queries when the database cannot fit
# entirely in memory. The only drawback of this setting is that longer checkpoint times
# may lead to slightly longer recovery times in case of a database or system crash.
# A lower number means lower IO pressure, and consequently longer checkpoint times.
# The configuration can also be commented out to remove the limitation entirely, and
# let the checkpointer flush data as fast as the hardware will go.
# Set this to -1 to disable the IOPS limit.
# dbms.checkpoint.iops.limit=300

# Enable a remote shell server which Neo4j Shell clients can log in to.
#dbms.shell.enabled=true
# The network interface IP the shell will listen on (use 0.0.0.0 for all interfaces).
#dbms.shell.host={{ node_ip_address }}
# The port the shell will listen on, default is 1337.
#dbms.shell.port=1337

# Only allow read operations from this Neo4j instance. This mode still requires
# write access to the directory for lock purposes.
#dbms.read_only=false

# Comma separated list of JAX-RS packages containing JAX-RS resources, one
# package name for each mountpoint. The listed package names will be loaded
# under the mountpoints specified. Uncomment this line to mount the
# org.neo4j.examples.server.unmanaged.HelloWorldResource.java from
# neo4j-server-examples under /examples/unmanaged, resulting in a final URL of
# http://localhost:7474/examples/unmanaged/helloworld/{nodeId}
#dbms.unmanaged_extension_classes=org.neo4j.examples.server.unmanaged=/examples/unmanaged

# Specified comma separated list of id types (like node or relationship) that should be reused.
# When some type is specified database will try to reuse corresponding ids as soon as it will be safe to do so.
# Currently only 'node' and 'relationship' types are supported.
# This settings is ignored in Neo4j Community Edition.
#dbms.ids.reuse.types.override=node,relationship

# Default values for the low-level graph engine
#neostore.nodestore.db.mapped_memory={{ neo4j_mm_nodestore }}
#neostore.relationshipstore.db.mapped_memory={{ neo4j_mm_relationshipstore }}
#neostore.propertystore.db.mapped_memory={{ neo4j_mm_propertystore }}
#neostore.propertystore.db.strings.mapped_memory={{ neo4j_mm_propertystore_strings }}
#neostore.propertystore.db.arrays.mapped_memory={{ neo4j_mm_propertystore_arrays }}
#neostore.propertystore.db.index.keys.mapped_memory={{ neo4j_nm_index_keys }}
#neostore.propertystore.db.index.mapped_memory={{ neo4j_mm_index }}
#cache_type={{ neo4j_cache_type }}

#********************************************************************
# JVM Parameters
#********************************************************************

# G1GC generally strikes a good balance between throughput and tail
# latency, without too much tuning.
dbms.jvm.additional=-XX:+UseG1GC

# Have common exceptions keep producing stack traces, so they can be
# debugged regardless of how often logs are rotated.
dbms.jvm.additional=-XX:-OmitStackTraceInFastThrow

# Make sure that `initmemory` is not only allocated, but committed to
# the process, before starting the database. This reduces memory
# fragmentation, increasing the effectiveness of transparent huge
# pages. It also reduces the possibility of seeing performance drop
# due to heap-growing GC events, where a decrease in available page
# cache leads to an increase in mean IO response time.
# Try reducing the heap memory, if this flag degrades performance.
dbms.jvm.additional=-XX:+AlwaysPreTouch

# Trust that non-static final fields are really final.
# This allows more optimizations and improves overall performance.
# NOTE: Disable this if you use embedded mode, or have extensions or dependencies that may use reflection or
# serialization to change the value of final fields!
dbms.jvm.additional=-XX:+UnlockExperimentalVMOptions
dbms.jvm.additional=-XX:+TrustFinalNonStaticFields

# Disable explicit garbage collection, which is occasionally invoked by the JDK itself.
dbms.jvm.additional=-XX:+DisableExplicitGC

# Remote JMX monitoring, uncomment and adjust the following lines as needed. Absolute paths to jmx.access and
# jmx.password files are required.
# Also make sure to update the jmx.access and jmx.password files with appropriate permission roles and passwords,
# the shipped configuration contains only a read only role called 'monitor' with password 'Neo4j'.
# For more details, see: http://download.oracle.com/javase/8/docs/technotes/guides/management/agent.html
# On Unix based systems the jmx.password file needs to be owned by the user that will run the server,
# and have permissions set to 0600.
# For details on setting these file permissions on Windows see:
#     http://docs.oracle.com/javase/8/docs/technotes/guides/management/security-windows.html
#dbms.jvm.additional=-Dcom.sun.management.jmxremote.port=3637
#dbms.jvm.additional=-Dcom.sun.management.jmxremote.authenticate=true
#dbms.jvm.additional=-Dcom.sun.management.jmxremote.ssl=false
#dbms.jvm.additional=-Dcom.sun.management.jmxremote.password.file=/absolute/path/to/conf/jmx.password
#dbms.jvm.additional=-Dcom.sun.management.jmxremote.access.file=/absolute/path/to/conf/jmx.access

# Some systems cannot discover host name automatically, and need this line configured:
#dbms.jvm.additional=-Djava.rmi.server.hostname=$THE_NEO4J_SERVER_HOSTNAME

# Expand Diffie Hellman (DH) key size from default 1024 to 2048 for DH-RSA cipher suites used in server TLS handshakes.
# This is to protect the server from any potential passive eavesdropping.
dbms.jvm.additional=-Djdk.tls.ephemeralDHKeySize=2048

# This mitigates a DDoS vector.
dbms.jvm.additional=-Djdk.tls.rejectClientInitiatedRenegotiation=true

#********************************************************************
# Wrapper Windows NT/2000/XP Service Properties
#********************************************************************
# WARNING - Do not modify any of these properties when an application
#  using this configuration file has been installed as a service.
#  Please uninstall the service before modifying this section.  The
#  service can then be reinstalled.

# Name of the service
dbms.windows_service_name=neo4j

#********************************************************************
# Other Neo4j system properties
#********************************************************************
dbms.jvm.additional=-Dunsupported.dbms.udc.source=tarball

dbms.security.procedures.unrestricted="apoc.*"
dbms.security.allow_csv_import_from_file_urls=true

#dbms.jvm.additional=-Dunsupported.dbms.udc.source=debian
#dbms.jvm.additional=-Dunsupported.dbms.udc.source=amazon-ami-CORE

#********************************************************************
# ES Neo4j
#********************************************************************

# This setting should only be set once for registering the framework and all the used submodules
dbms.unmanaged_extension_classes=com.graphaware.server=/graphaware

com.graphaware.runtime.enabled=true

#UIDM becomes the module ID:
com.graphaware.module.UIDM.1=com.graphaware.module.uuid.UuidBootstrapper

#optional, default is "uuid". (only if using the UUID module)
com.graphaware.module.UIDM.uuidProperty=_bldrId

#optional, default is all nodes:
#com.graphaware.module.UIDM.node=hasLabel('Label1') || hasLabel('Label2')

#optional, default is uuidIndex
com.graphaware.module.UIDM.uuidIndex=uuidIndex

#prevent the whole db to be assigned a new uuid if the uuid module is settle up together with neo4j2es
com.graphaware.module.UIDM.initializeUntil=1557511200000

#ES becomes the module ID:
com.graphaware.module.ES.2=com.graphaware.module.es.ElasticSearchModuleBootstrapper

#URI of Elasticsearch
com.graphaware.module.ES.uri=172.27.0.3

#Port of Elasticsearch
com.graphaware.module.ES.port=9200

#optional, protocol of Elasticsearch connection, defaults to http
com.graphaware.module.ES.protocol=http

#optional, Elasticsearch index name, default is neo4j-index
#com.graphaware.module.ES.index=bldr_data_idx

#optional, node property key of a propery that is used as unique identifier of the node. Must be the same as com.graphaware.module.UIDM.uuidProperty (only if using UUID module), defaults to uuid
#use "ID()" to use native Neo4j IDs as Elasticsearch IDs (not recommended)
#com.graphaware.module.ES.keyProperty=_bldrId

#optional, whether to retry if a replication fails, defaults to false
com.graphaware.module.ES.retryOnError=true

#optional, size of the in-memory queue that queues up operations to be synchronised to Elasticsearch, defaults to 10000
com.graphaware.module.ES.queueSize=10000

#optional, size of the batch size to use during re-initialization, defaults to 1000
com.graphaware.module.ES.reindexBatchSize=200

#optional, specify which nodes to index in Elasticsearch, defaults to all nodes
#com.graphaware.module.ES.node=hasLabel('context')

#optional, specify which node properties to index in Elasticsearch, defaults to all properties
#com.graphaware.module.ES.node.property=key != 'age'

#optional, specify whether to send updates to Elasticsearch in bulk, defaults to true (highly recommended)
com.graphaware.module.ES.bulk=false

#optional, read explanation below, defaults to 0
com.graphaware.module.ES.initializeUntil=1557511200000

#optional, whether or not the reindexation process (when db start) should be made in asynchronous mode
#default is "false" and the db will not be available until completed
#com.graphaware.module.ES.asyncIndexation=true

#####################################################################
# From Team
#####################################################################
#com.graphaware.runtime.enabled=true
#com.graphaware.module.ES.1=com.graphaware.module.es.ElasticSearchModuleBootstrapper
#com.graphaware.module.ES.uri=http://localhost:9200
#com.graphaware.module.ES.port=443
#com.graphaware.module.ES.protocol=https
com.graphaware.module.ES.keyProperty=_bldrId
com.graphaware.module.ES.relationship=(true)
com.graphaware.module.ES.mapping=com.graphaware.module.es.mapping.JsonFileMapping
com.graphaware.module.ES.file=mapping.json
#com.graphaware.module.ES.bulk=true
#com.graphaware.module.ES.queueSize=100

#####################################################################   
# UUID EXTRA 
#####################################################################

#optional, default is no relationships:
#com.graphaware.module.UIDM.relationship=isType('Type1')

#optional, default is uuidRelIndex
com.graphaware.module.UIDM.uuidRelationshipIndex=uuidRelIndex

#assign uuid's to all relationships
com.graphaware.module.UIDM.relationship=com.graphaware.runtime.policy.all.IncludeAllBusinessRelationships

debug.log

core2_1  | Active database: graph.db
core2_1  | Directories in use:
core2_1  |   home:         /var/lib/neo4j
core2_1  |   config:       /var/lib/neo4j/conf
core2_1  |   logs:         /var/lib/neo4j/logs
core2_1  |   plugins:      /var/lib/neo4j/plugins
core2_1  |   import:       /var/lib/neo4j/import
core2_1  |   data:         /var/lib/neo4j/data
core2_1  |   certificates: /var/lib/neo4j/certificates
core2_1  |   run:          /var/lib/neo4j/run
core2_1  | Starting Neo4j.
core3_1  | Active database: graph.db
core3_1  | Directories in use:
core3_1  |   home:         /var/lib/neo4j
core3_1  |   config:       /var/lib/neo4j/conf
core3_1  |   logs:         /var/lib/neo4j/logs
core3_1  |   plugins:      /var/lib/neo4j/plugins
core3_1  |   import:       /var/lib/neo4j/import
core3_1  |   data:         /var/lib/neo4j/data
core3_1  |   certificates: /var/lib/neo4j/certificates
core3_1  |   run:          /var/lib/neo4j/run
core3_1  | Starting Neo4j.
core2_1  | 2019-05-04 10:12:25.379+0000 WARN  ha.host.data is deprecated.
core2_1  | 2019-05-04 10:12:25.401+0000 INFO  ======== Neo4j 3.5.1 ========
core2_1  | 2019-05-04 10:12:25.406+0000 INFO  Starting...
core-placeholder | Active database: graph.db
core-placeholder | Directories in use:
core-placeholder |   home:         /var/lib/neo4j
core-placeholder |   config:       /var/lib/neo4j/conf
core-placeholder |   logs:         /var/lib/neo4j/logs
core-placeholder |   plugins:      /var/lib/neo4j/plugins
core-placeholder |   import:       /var/lib/neo4j/import
core-placeholder |   data:         /var/lib/neo4j/data
core-placeholder |   certificates: /var/lib/neo4j/certificates
core-placeholder |   run:          /var/lib/neo4j/run
core-placeholder | Starting Neo4j.
core3_1  | 2019-05-04 10:12:26.819+0000 WARN  ha.host.data is deprecated.
core3_1  | 2019-05-04 10:12:26.847+0000 INFO  ======== Neo4j 3.5.1 ========
core3_1  | 2019-05-04 10:12:26.853+0000 INFO  Starting...
core2_1  | 2019-05-04 10:12:27.375+0000 INFO  Initiating metrics...
core2_1  | 2019-05-04 10:12:27.482+0000 INFO  My connection info: [
core2_1  |  Discovery:   listen=0.0.0.0:5000, advertised=f47a733d0b70:5000,
core2_1  |  Transaction: listen=0.0.0.0:6000, advertised=f47a733d0b70:6000, 
core2_1  |  Raft:        listen=0.0.0.0:7000, advertised=f47a733d0b70:7000, 
core2_1  |  Client Connector Addresses: bolt://localhost:7689,http://localhost:7476,https://localhost:6478
core2_1  | ]
core2_1  | 2019-05-04 10:12:27.483+0000 INFO  Discovering other core members in initial members set: [core1:5000, core2:5000, core3:5000]
core-placeholder | 2019-05-04 10:12:27.825+0000 WARN  ha.host.data is deprecated.
core-placeholder | 2019-05-04 10:12:27.873+0000 INFO  ======== Neo4j 3.5.1 ========
core-placeholder | 2019-05-04 10:12:27.883+0000 INFO  Starting...
core3_1  | 2019-05-04 10:12:29.321+0000 INFO  Initiating metrics...
core3_1  | 2019-05-04 10:12:29.404+0000 INFO  My connection info: [
core3_1  |  Discovery:   listen=0.0.0.0:5000, advertised=core3:5000,
core3_1  |  Transaction: listen=0.0.0.0:6000, advertised=core3:6000, 
core3_1  |  Raft:        listen=0.0.0.0:7000, advertised=core3:7000, 
core3_1  |  Client Connector Addresses: bolt://localhost:7690,http://localhost:7477,https://localhost:6479
core3_1  | ]
core3_1  | 2019-05-04 10:12:29.408+0000 INFO  Discovering other core members in initial members set: [core1:5000, core2:5000, core3:5000]
core-placeholder | 2019-05-04 10:12:30.105+0000 INFO  Initiating metrics...
core-placeholder | 2019-05-04 10:12:30.181+0000 INFO  My connection info: [
core-placeholder |  Discovery:   listen=0.0.0.0:5000, advertised=2a2b1eed9f80:5000,
core-placeholder |  Transaction: listen=0.0.0.0:6000, advertised=2a2b1eed9f80:6000, 
core-placeholder |  Raft:        listen=0.0.0.0:7000, advertised=2a2b1eed9f80:7000, 
core-placeholder |  Client Connector Addresses: bolt://localhost:7688,http://localhost:7475,https://localhost:6477
core-placeholder | ]
core-placeholder | 2019-05-04 10:12:30.182+0000 INFO  Discovering other core members in initial members set: [core1:5000, core2:5000, core3:5000]
core2_1  | 2019-05-04 10:12:34.862+0000 INFO  Discovered core member at core3:5000
core2_1  | 2019-05-04 10:12:34.903+0000 INFO  Bound to cluster with id f3375433-757f-4e57-88a4-12e27c74d383
core2_1  | 2019-05-04 10:12:34.904+0000 INFO  Discovered core member at core1:5000
core2_1  | 2019-05-04 10:12:36.094+0000 INFO  GraphAware Runtime enabled, bootstrapping...
core2_1  | 2019-05-04 10:12:36.120+0000 INFO  onCoreTopologyChange {clusterId=ClusterId{uuid=f3375433-757f-4e57-88a4-12e27c74d383}, bootstrappable=true, coreMembers={MemberId{2b63eb1f}=CoreServerInfo{raftServer=core3:7000, catchupServer=core3:6000, clientConnectorAddresses=bolt://localhost:7690,http://localhost:7477,https://localhost:6479, groups=[], database=default, refuseToBeLeader=false}, MemberId{555f97a8}=CoreServerInfo{raftServer=2a2b1eed9f80:7000, catchupServer=2a2b1eed9f80:6000, clientConnectorAddresses=bolt://localhost:7688,http://localhost:7475,https://localhost:6477, groups=[], database=default, refuseToBeLeader=false}, MemberId{44524f83}=CoreServerInfo{raftServer=f47a733d0b70:7000, catchupServer=f47a733d0b70:6000, clientConnectorAddresses=bolt://localhost:7689,http://localhost:7476,https://localhost:6478, groups=[], database=default, refuseToBeLeader=false}}}
core2_1  | 2019-05-04 10:12:36.123+0000 INFO  Bootstrapping module with order 1, ID UIDM, using com.graphaware.module.uuid.UuidBootstrapper
core2_1  | 2019-05-04 10:12:36.196+0000 INFO  UIDM (re-)initialize until set to 1557511200000
core2_1  | 2019-05-04 10:12:36.196+0000 INFO  That's 546443804 ms in the future
core2_1  | 2019-05-04 10:12:36.201+0000 INFO  Relationship Inclusion Policy set to com.graphaware.common.policy.inclusion.composite.CompositeRelationshipInclusionPolicy@355a19e1
core2_1  | 2019-05-04 10:12:36.201+0000 INFO  uuidProperty set to _bldrId
core2_1  | 2019-05-04 10:12:36.201+0000 INFO  uuidIndex set to uuidIndex
core2_1  | 2019-05-04 10:12:36.201+0000 INFO  uuidRelationshipIndex set to uuidRelIndex
core2_1  | 2019-05-04 10:12:36.210+0000 INFO  Registering module UIDM with GraphAware Runtime.
core2_1  | 2019-05-04 10:12:36.211+0000 INFO  Bootstrapping module with order 2, ID ES, using com.graphaware.module.es.ElasticSearchModuleBootstrapper
core2_1  | 2019-05-04 10:12:36.220+0000 INFO  ElasticSearch index-prefix set to neo4j-index
core2_1  | 2019-05-04 10:12:36.220+0000 INFO  ElasticSearch key-property set to uuid
core2_1  | 2019-05-04 10:12:36.220+0000 INFO  ElasticSearch force-strings set to false
core2_1  | 2019-05-04 10:12:36.220+0000 INFO  ES (re-)initialize until set to 1557511200000
core2_1  | 2019-05-04 10:12:36.221+0000 INFO  That's 546443779 ms in the future
core2_1  | 2019-05-04 10:12:36.266+0000 INFO  Relationship Inclusion Policy set to com.graphaware.common.policy.inclusion.composite.CompositeRelationshipInclusionPolicy@7e895e63
core2_1  | 2019-05-04 10:12:36.267+0000 INFO  Elasticsearch URI set to 172.27.0.3
core2_1  | 2019-05-04 10:12:36.267+0000 INFO  Elasticsearch protocol set to http
core2_1  | 2019-05-04 10:12:36.267+0000 INFO  Elasticsearch port set to 9200
core2_1  | 2019-05-04 10:12:36.267+0000 INFO  Elasticsearch key property set to _bldrId
core2_1  | 2019-05-04 10:12:36.267+0000 INFO  Elasticsearch retry-on-error set to true
core2_1  | 2019-05-04 10:12:36.268+0000 INFO  Elasticsearch module queue capacity set to 10000
core2_1  | 2019-05-04 10:12:36.268+0000 INFO  Elasticsearch module reindex batch size set to 200
core2_1  | 2019-05-04 10:12:36.268+0000 INFO  Elasticsearch bulk execution set to false
core2_1  | 2019-05-04 10:12:36.276+0000 INFO  Using mapping file at path /var/lib/neo4j/conf/mapping.json
core2_1  | 2019-05-04 10:12:36.581+0000 INFO  Elasticsearch mapping configured with class com.graphaware.module.es.mapping.JsonFileMapping
core2_1  | 2019-05-04 10:12:36.600+0000 INFO  Registering module ES with GraphAware Runtime.
core2_1  | 2019-05-04 10:12:36.601+0000 INFO  GraphAware Runtime bootstrapped, starting the Runtime...
core2_1  | 2019-05-04 10:12:36.885+0000 INFO  Starting GraphAware...
core2_1  | 2019-05-04 10:12:36.887+0000 INFO  Loading module metadata...
core2_1  | 2019-05-04 10:12:36.888+0000 INFO  Loading metadata for module UIDM
core2_1  | 2019-05-04 10:12:37.034+0000 INFO  Module UIDM seems to have been registered before, metadata loaded successfully.
core2_1  | 2019-05-04 10:12:37.035+0000 INFO  Module UIDM seems to have changed configuration since last run, will try to re-initialize...
core2_1  | 2019-05-04 10:12:37.036+0000 INFO  Instance not writable. Will NOT re-initialize.
core2_1  | 2019-05-04 10:12:37.037+0000 INFO  Loading metadata for module ES
core2_1  | 2019-05-04 10:12:37.052+0000 ERROR Could not deserialize metadata for module ID ES
core2_1  | 2019-05-04 10:12:37.053+0000 INFO  Module ES seems to have corrupted metadata.
core2_1  | 2019-05-04 10:12:37.053+0000 INFO  Module ES seems to have corrupted metadata, will try to re-initialize...
core2_1  | 2019-05-04 10:12:37.053+0000 INFO  Instance not writable. Will NOT re-initialize.
core2_1  | 2019-05-04 10:12:37.054+0000 INFO  Creating fresh metadata for module ES.
core2_1  | 2019-05-04 10:12:37.054+0000 INFO  Module ES has not changed configuration since last run, already initialized.
core2_1  | 2019-05-04 10:12:37.056+0000 INFO  Module metadata loaded.
core2_1  | 2019-05-04 10:12:37.057+0000 INFO  Starting transaction-driven modules...
core2_1  | 2019-05-04 10:12:37.057+0000 INFO  Starting Elasticsearch Writer...
core2_1  | 2019-05-04 10:12:37.090+0000 INFO  Creating Jest Client...
core2_1  | 2019-05-04 10:12:37.314+0000 INFO  Created Jest Client.
core2_1  | 2019-05-04 10:12:37.469+0000 INFO  Started Elasticsearch Writer.
core2_1  | 2019-05-04 10:12:37.470+0000 INFO  Transaction-driven modules started.
core2_1  | 2019-05-04 10:12:37.470+0000 INFO  There are no timer-driven runtime modules. Not scheduling any tasks.
core2_1  | 2019-05-04 10:12:37.470+0000 INFO  GraphAware started.
core2_1  | 2019-05-04 10:12:37.470+0000 INFO  GraphAware Runtime automatically started.
core-placeholder | 2019-05-04 10:12:38.619+0000 INFO  Bound to cluster with id f3375433-757f-4e57-88a4-12e27c74d383
core-placeholder | 2019-05-04 10:12:38.638+0000 INFO  Discovered core member at core2:5000
core-placeholder | 2019-05-04 10:12:38.675+0000 INFO  Discovered core member at core3:5000
core3_1  | 2019-05-04 10:12:38.910+0000 INFO  Bound to cluster with id f3375433-757f-4e57-88a4-12e27c74d383
core3_1  | 2019-05-04 10:12:39.001+0000 INFO  Discovered core member at core2:5000
core3_1  | 2019-05-04 10:12:39.038+0000 INFO  Discovered core member at core1:5000
core-placeholder | 2019-05-04 10:12:40.281+0000 INFO  GraphAware Runtime enabled, bootstrapping...
core-placeholder | 2019-05-04 10:12:40.321+0000 INFO  onCoreTopologyChange {clusterId=ClusterId{uuid=f3375433-757f-4e57-88a4-12e27c74d383}, bootstrappable=false, coreMembers={MemberId{2b63eb1f}=CoreServerInfo{raftServer=core3:7000, catchupServer=core3:6000, clientConnectorAddresses=bolt://localhost:7690,http://localhost:7477,https://localhost:6479, groups=[], database=default, refuseToBeLeader=false}, MemberId{555f97a8}=CoreServerInfo{raftServer=2a2b1eed9f80:7000, catchupServer=2a2b1eed9f80:6000, clientConnectorAddresses=bolt://localhost:7688,http://localhost:7475,https://localhost:6477, groups=[], database=default, refuseToBeLeader=false}, MemberId{44524f83}=CoreServerInfo{raftServer=f47a733d0b70:7000, catchupServer=f47a733d0b70:6000, clientConnectorAddresses=bolt://localhost:7689,http://localhost:7476,https://localhost:6478, groups=[], database=default, refuseToBeLeader=false}}}
core-placeholder | 2019-05-04 10:12:40.327+0000 INFO  Bootstrapping module with order 1, ID UIDM, using com.graphaware.module.uuid.UuidBootstrapper
core3_1  | 2019-05-04 10:12:40.435+0000 INFO  GraphAware Runtime enabled, bootstrapping...
core-placeholder | 2019-05-04 10:12:40.448+0000 INFO  UIDM (re-)initialize until set to 1557511200000
core-placeholder | 2019-05-04 10:12:40.448+0000 INFO  That's 546439552 ms in the future
core-placeholder | 2019-05-04 10:12:40.456+0000 INFO  Relationship Inclusion Policy set to com.graphaware.common.policy.inclusion.composite.CompositeRelationshipInclusionPolicy@5ff3c3a1
core-placeholder | 2019-05-04 10:12:40.457+0000 INFO  uuidProperty set to _bldrId
core-placeholder | 2019-05-04 10:12:40.457+0000 INFO  uuidIndex set to uuidIndex
core-placeholder | 2019-05-04 10:12:40.458+0000 INFO  uuidRelationshipIndex set to uuidRelIndex
core-placeholder | 2019-05-04 10:12:40.473+0000 INFO  Registering module UIDM with GraphAware Runtime.
core-placeholder | 2019-05-04 10:12:40.475+0000 INFO  Bootstrapping module with order 2, ID ES, using com.graphaware.module.es.ElasticSearchModuleBootstrapper
core-placeholder | 2019-05-04 10:12:40.490+0000 INFO  ElasticSearch index-prefix set to neo4j-index
core-placeholder | 2019-05-04 10:12:40.491+0000 INFO  ElasticSearch key-property set to uuid
core-placeholder | 2019-05-04 10:12:40.491+0000 INFO  ElasticSearch force-strings set to false
core-placeholder | 2019-05-04 10:12:40.492+0000 INFO  ES (re-)initialize until set to 1557511200000
core-placeholder | 2019-05-04 10:12:40.493+0000 INFO  That's 546439507 ms in the future
core3_1  | 2019-05-04 10:12:40.503+0000 INFO  onCoreTopologyChange {clusterId=ClusterId{uuid=f3375433-757f-4e57-88a4-12e27c74d383}, bootstrappable=false, coreMembers={MemberId{2b63eb1f}=CoreServerInfo{raftServer=core3:7000, catchupServer=core3:6000, clientConnectorAddresses=bolt://localhost:7690,http://localhost:7477,https://localhost:6479, groups=[], database=default, refuseToBeLeader=false}, MemberId{555f97a8}=CoreServerInfo{raftServer=2a2b1eed9f80:7000, catchupServer=2a2b1eed9f80:6000, clientConnectorAddresses=bolt://localhost:7688,http://localhost:7475,https://localhost:6477, groups=[], database=default, refuseToBeLeader=false}, MemberId{44524f83}=CoreServerInfo{raftServer=f47a733d0b70:7000, catchupServer=f47a733d0b70:6000, clientConnectorAddresses=bolt://localhost:7689,http://localhost:7476,https://localhost:6478, groups=[], database=default, refuseToBeLeader=false}}}
core3_1  | 2019-05-04 10:12:40.510+0000 INFO  Bootstrapping module with order 1, ID UIDM, using com.graphaware.module.uuid.UuidBootstrapper
core-placeholder | 2019-05-04 10:12:40.566+0000 INFO  Relationship Inclusion Policy set to com.graphaware.common.policy.inclusion.composite.CompositeRelationshipInclusionPolicy@7fce3ad5
core-placeholder | 2019-05-04 10:12:40.566+0000 INFO  Elasticsearch URI set to 172.27.0.3
core-placeholder | 2019-05-04 10:12:40.566+0000 INFO  Elasticsearch protocol set to http
core-placeholder | 2019-05-04 10:12:40.567+0000 INFO  Elasticsearch port set to 9200
core-placeholder | 2019-05-04 10:12:40.567+0000 INFO  Elasticsearch key property set to _bldrId
core-placeholder | 2019-05-04 10:12:40.567+0000 INFO  Elasticsearch retry-on-error set to true
core-placeholder | 2019-05-04 10:12:40.567+0000 INFO  Elasticsearch module queue capacity set to 10000
core-placeholder | 2019-05-04 10:12:40.568+0000 INFO  Elasticsearch module reindex batch size set to 200
core-placeholder | 2019-05-04 10:12:40.569+0000 INFO  Elasticsearch bulk execution set to false
core-placeholder | 2019-05-04 10:12:40.579+0000 INFO  Using mapping file at path /var/lib/neo4j/conf/mapping.json
core3_1  | 2019-05-04 10:12:40.633+0000 INFO  UIDM (re-)initialize until set to 1557511200000
core3_1  | 2019-05-04 10:12:40.634+0000 INFO  That's 546439366 ms in the future
core3_1  | 2019-05-04 10:12:40.642+0000 INFO  Relationship Inclusion Policy set to com.graphaware.common.policy.inclusion.composite.CompositeRelationshipInclusionPolicy@6cbb4241
core3_1  | 2019-05-04 10:12:40.642+0000 INFO  uuidProperty set to _bldrId
core3_1  | 2019-05-04 10:12:40.643+0000 INFO  uuidIndex set to uuidIndex
core3_1  | 2019-05-04 10:12:40.643+0000 INFO  uuidRelationshipIndex set to uuidRelIndex
core3_1  | 2019-05-04 10:12:40.657+0000 INFO  Registering module UIDM with GraphAware Runtime.
core3_1  | 2019-05-04 10:12:40.659+0000 INFO  Bootstrapping module with order 2, ID ES, using com.graphaware.module.es.ElasticSearchModuleBootstrapper
core3_1  | 2019-05-04 10:12:40.673+0000 INFO  ElasticSearch index-prefix set to neo4j-index
core3_1  | 2019-05-04 10:12:40.674+0000 INFO  ElasticSearch key-property set to uuid
core3_1  | 2019-05-04 10:12:40.674+0000 INFO  ElasticSearch force-strings set to false
core3_1  | 2019-05-04 10:12:40.675+0000 INFO  ES (re-)initialize until set to 1557511200000
core3_1  | 2019-05-04 10:12:40.675+0000 INFO  That's 546439325 ms in the future
core3_1  | 2019-05-04 10:12:40.746+0000 INFO  Relationship Inclusion Policy set to com.graphaware.common.policy.inclusion.composite.CompositeRelationshipInclusionPolicy@6c2f7d80
core3_1  | 2019-05-04 10:12:40.747+0000 INFO  Elasticsearch URI set to 172.27.0.3
core3_1  | 2019-05-04 10:12:40.748+0000 INFO  Elasticsearch protocol set to http
core3_1  | 2019-05-04 10:12:40.748+0000 INFO  Elasticsearch port set to 9200
core3_1  | 2019-05-04 10:12:40.749+0000 INFO  Elasticsearch key property set to _bldrId
core3_1  | 2019-05-04 10:12:40.749+0000 INFO  Elasticsearch retry-on-error set to true
core3_1  | 2019-05-04 10:12:40.750+0000 INFO  Elasticsearch module queue capacity set to 10000
core3_1  | 2019-05-04 10:12:40.751+0000 INFO  Elasticsearch module reindex batch size set to 200
core3_1  | 2019-05-04 10:12:40.752+0000 INFO  Elasticsearch bulk execution set to false
core3_1  | 2019-05-04 10:12:40.765+0000 INFO  Using mapping file at path /var/lib/neo4j/conf/mapping.json
core-placeholder | 2019-05-04 10:12:41.051+0000 INFO  Elasticsearch mapping configured with class com.graphaware.module.es.mapping.JsonFileMapping
core-placeholder | 2019-05-04 10:12:41.074+0000 INFO  Registering module ES with GraphAware Runtime.
core-placeholder | 2019-05-04 10:12:41.074+0000 INFO  GraphAware Runtime bootstrapped, starting the Runtime...
core3_1  | 2019-05-04 10:12:41.209+0000 INFO  Elasticsearch mapping configured with class com.graphaware.module.es.mapping.JsonFileMapping
core3_1  | 2019-05-04 10:12:41.227+0000 INFO  Registering module ES with GraphAware Runtime.
core3_1  | 2019-05-04 10:12:41.228+0000 INFO  GraphAware Runtime bootstrapped, starting the Runtime...
core-placeholder | 2019-05-04 10:12:41.673+0000 INFO  Starting GraphAware...
core-placeholder | 2019-05-04 10:12:41.687+0000 INFO  Loading module metadata...
core-placeholder | 2019-05-04 10:12:41.688+0000 INFO  Loading metadata for module UIDM
core3_1  | 2019-05-04 10:12:41.764+0000 INFO  Starting GraphAware...
core3_1  | 2019-05-04 10:12:41.767+0000 INFO  Loading module metadata...
core3_1  | 2019-05-04 10:12:41.768+0000 INFO  Loading metadata for module UIDM
core-placeholder | 2019-05-04 10:12:41.872+0000 INFO  Module UIDM seems to have been registered before, metadata loaded successfully.
core-placeholder | 2019-05-04 10:12:41.872+0000 INFO  Module UIDM seems to have changed configuration since last run, will try to re-initialize...
core-placeholder | 2019-05-04 10:12:41.873+0000 INFO  Instance not writable. Will NOT re-initialize.
core-placeholder | 2019-05-04 10:12:41.873+0000 INFO  Loading metadata for module ES
core-placeholder | 2019-05-04 10:12:41.903+0000 ERROR Could not deserialize metadata for module ID ES
core-placeholder | 2019-05-04 10:12:41.910+0000 INFO  Module ES seems to have corrupted metadata.
core-placeholder | 2019-05-04 10:12:41.910+0000 INFO  Module ES seems to have corrupted metadata, will try to re-initialize...
core-placeholder | 2019-05-04 10:12:41.911+0000 INFO  Instance not writable. Will NOT re-initialize.
core-placeholder | 2019-05-04 10:12:41.911+0000 INFO  Creating fresh metadata for module ES.
core-placeholder | 2019-05-04 10:12:41.911+0000 INFO  Module ES has not changed configuration since last run, already initialized.
core-placeholder | 2019-05-04 10:12:41.913+0000 INFO  Module metadata loaded.
core-placeholder | 2019-05-04 10:12:41.913+0000 INFO  Starting transaction-driven modules...
core-placeholder | 2019-05-04 10:12:41.913+0000 INFO  Starting Elasticsearch Writer...
core3_1  | 2019-05-04 10:12:41.924+0000 INFO  Module UIDM seems to have been registered before, metadata loaded successfully.
core3_1  | 2019-05-04 10:12:41.924+0000 INFO  Module UIDM seems to have changed configuration since last run, will try to re-initialize...
core3_1  | 2019-05-04 10:12:41.926+0000 INFO  Instance not writable. Will NOT re-initialize.
core3_1  | 2019-05-04 10:12:41.926+0000 INFO  Loading metadata for module ES
core3_1  | 2019-05-04 10:12:41.945+0000 ERROR Could not deserialize metadata for module ID ES
core3_1  | 2019-05-04 10:12:41.946+0000 INFO  Module ES seems to have corrupted metadata.
core3_1  | 2019-05-04 10:12:41.946+0000 INFO  Module ES seems to have corrupted metadata, will try to re-initialize...
core3_1  | 2019-05-04 10:12:41.947+0000 INFO  Instance not writable. Will NOT re-initialize.
core3_1  | 2019-05-04 10:12:41.947+0000 INFO  Creating fresh metadata for module ES.
core3_1  | 2019-05-04 10:12:41.947+0000 INFO  Module ES has not changed configuration since last run, already initialized.
core-placeholder | 2019-05-04 10:12:41.953+0000 INFO  Creating Jest Client...
core3_1  | 2019-05-04 10:12:41.957+0000 INFO  Module metadata loaded.
core3_1  | 2019-05-04 10:12:41.958+0000 INFO  Starting transaction-driven modules...
core3_1  | 2019-05-04 10:12:41.958+0000 INFO  Starting Elasticsearch Writer...
core3_1  | 2019-05-04 10:12:42.047+0000 INFO  Creating Jest Client...
core-placeholder | 2019-05-04 10:12:42.266+0000 INFO  Created Jest Client.
core3_1  | 2019-05-04 10:12:42.536+0000 INFO  Created Jest Client.
core-placeholder | 2019-05-04 10:12:42.656+0000 INFO  Started Elasticsearch Writer.
core-placeholder | 2019-05-04 10:12:42.656+0000 INFO  Transaction-driven modules started.
core-placeholder | 2019-05-04 10:12:42.656+0000 INFO  There are no timer-driven runtime modules. Not scheduling any tasks.
core-placeholder | 2019-05-04 10:12:42.656+0000 INFO  GraphAware started.
core-placeholder | 2019-05-04 10:12:42.657+0000 INFO  GraphAware Runtime automatically started.
core3_1  | 2019-05-04 10:12:42.866+0000 INFO  Started Elasticsearch Writer.
core3_1  | 2019-05-04 10:12:42.867+0000 INFO  Transaction-driven modules started.
core3_1  | 2019-05-04 10:12:42.867+0000 INFO  There are no timer-driven runtime modules. Not scheduling any tasks.
core3_1  | 2019-05-04 10:12:42.868+0000 INFO  GraphAware started.
core3_1  | 2019-05-04 10:12:42.868+0000 INFO  GraphAware Runtime automatically started.
core2_1  | 2019-05-04 10:12:51.058+0000 INFO  Connected to core3/172.27.0.7:7000 [raft version:2]
core2_1  | 2019-05-04 10:12:51.058+0000 INFO  Connected to 2a2b1eed9f80/172.27.0.8:7000 [raft version:2]
core3_1  | 2019-05-04 10:12:51.249+0000 INFO  Connected to f47a733d0b70/172.27.0.6:7000 [raft version:2]
core-placeholder | 2019-05-04 10:12:51.251+0000 INFO  Connected to f47a733d0b70/172.27.0.6:7000 [raft version:2]
core2_1  | 2019-05-04 10:13:05.414+0000 INFO  Waiting to catchup with leader... we are 0 entries behind leader at 36.
core2_1  | 2019-05-04 10:13:05.415+0000 INFO  Successfully joined the Raft group.
core3_1  | 2019-05-04 10:13:10.738+0000 INFO  Waiting to catchup with leader... we are 0 entries behind leader at 36.
core3_1  | 2019-05-04 10:13:10.738+0000 INFO  Successfully joined the Raft group.
core-placeholder | 2019-05-04 10:13:10.785+0000 INFO  Waiting to catchup with leader... we are 0 entries behind leader at 36.
core-placeholder | 2019-05-04 10:13:10.785+0000 INFO  Successfully joined the Raft group.
core2_1  | 2019-05-04 10:13:15.277+0000 INFO  Sending metrics to CSV file at /var/lib/neo4j/metrics
core2_1  | 2019-05-04 10:13:16.149+0000 INFO  Bolt enabled on 0.0.0.0:7689.
core2_1  | 2019-05-04 10:13:18.298+0000 WARN  Server thread metrics not available (missing neo4j.server.threads.jetty.all)
core2_1  | 2019-05-04 10:13:18.299+0000 WARN  Server thread metrics not available (missing neo4j.server.threads.jetty.idle)
core2_1  | 2019-05-04 10:13:20.333+0000 INFO  Started.
core2_1  | 2019-05-04 10:13:20.608+0000 INFO  Mounted REST API at: /db/manage
core2_1  | 2019-05-04 10:13:20.635+0000 INFO  started
core2_1  | 2019-05-04 10:13:20.638+0000 INFO  Mounted unmanaged extension [com.graphaware.server] at [/graphaware]
core2_1  | 2019-05-04 10:13:20.748+0000 INFO  Server thread metrics have been registered successfully
core2_1  | 2019-05-04 10:13:20.948+0000 INFO  Google Analytics enabled
core2_1  | 2019-05-04 10:13:20.958+0000 INFO  Mounting GraphAware Framework at /graphaware
core2_1  | 2019-05-04 10:13:20.977+0000 INFO  Will try to scan the following packages: {com.**.graphaware.**,org.**.graphaware.**,net.**.graphaware.**}
core2_1  | 2019-05-04 10:13:24.639+0000 INFO  Remote interface available at http://localhost:7476/
core-placeholder | 2019-05-04 10:13:25.011+0000 INFO  Sending metrics to CSV file at /var/lib/neo4j/metrics
core3_1  | 2019-05-04 10:13:25.206+0000 INFO  Sending metrics to CSV file at /var/lib/neo4j/metrics
core-placeholder | 2019-05-04 10:13:25.732+0000 INFO  Bolt enabled on 0.0.0.0:7688.
core3_1  | 2019-05-04 10:13:26.241+0000 INFO  Bolt enabled on 0.0.0.0:7690.
core-placeholder | 2019-05-04 10:13:28.018+0000 INFO  Started.
core-placeholder | 2019-05-04 10:13:28.024+0000 WARN  Server thread metrics not available (missing neo4j.server.threads.jetty.all)
core-placeholder | 2019-05-04 10:13:28.025+0000 WARN  Server thread metrics not available (missing neo4j.server.threads.jetty.idle)
core3_1  | 2019-05-04 10:13:28.218+0000 WARN  Server thread metrics not available (missing neo4j.server.threads.jetty.all)
core3_1  | 2019-05-04 10:13:28.219+0000 WARN  Server thread metrics not available (missing neo4j.server.threads.jetty.idle)
core-placeholder | 2019-05-04 10:13:28.292+0000 INFO  Mounted REST API at: /db/manage
core-placeholder | 2019-05-04 10:13:28.311+0000 INFO  started
core-placeholder | 2019-05-04 10:13:28.313+0000 INFO  Mounted unmanaged extension [com.graphaware.server] at [/graphaware]
core-placeholder | 2019-05-04 10:13:28.464+0000 INFO  Server thread metrics have been registered successfully
core-placeholder | 2019-05-04 10:13:28.603+0000 INFO  Google Analytics enabled
core3_1  | 2019-05-04 10:13:28.610+0000 INFO  Started.
core-placeholder | 2019-05-04 10:13:28.611+0000 INFO  Mounting GraphAware Framework at /graphaware
core-placeholder | 2019-05-04 10:13:28.617+0000 INFO  Will try to scan the following packages: {com.**.graphaware.**,org.**.graphaware.**,net.**.graphaware.**}
core3_1  | 2019-05-04 10:13:28.900+0000 INFO  Mounted REST API at: /db/manage
core3_1  | 2019-05-04 10:13:28.919+0000 INFO  started
core3_1  | 2019-05-04 10:13:28.921+0000 INFO  Mounted unmanaged extension [com.graphaware.server] at [/graphaware]
core3_1  | 2019-05-04 10:13:28.997+0000 INFO  Server thread metrics have been registered successfully
core3_1  | 2019-05-04 10:13:29.119+0000 INFO  Google Analytics enabled
core3_1  | 2019-05-04 10:13:29.126+0000 INFO  Mounting GraphAware Framework at /graphaware
core3_1  | 2019-05-04 10:13:29.131+0000 INFO  Will try to scan the following packages: {com.**.graphaware.**,org.**.graphaware.**,net.**.graphaware.**}
core-placeholder | 2019-05-04 10:13:31.194+0000 INFO  Remote interface available at http://localhost:7475/
core3_1  | 2019-05-04 10:13:31.711+0000 INFO  Remote interface available at http://localhost:7477/
itaybraha commented 5 years ago

@ikwattro we are using graphaware-neo4j-to-elasticsearch-3.5.1.53.11.jar in the database we are dumping from which is running Neo4j 3.5.1, as far as I can see that is the plugin matching the database version. Would you recommend upgrading the plugin and doing another dump or would this version of the plugin suffice?

ikwattro commented 5 years ago

Hmm no, if possible :

  1. Load your backup with a neo4j in single mode
  2. Delete all nodes and relationships so there is no sensitive data
  3. Make a backup again
  4. Send me the empty database backup ( but that contains the metadata )