JanusGraph / janusgraph

JanusGraph: an open-source, distributed graph database
https://janusgraph.org
Other
5.3k stars 1.17k forks source link

Unique index failed when vertex creation running in parallel #121

Open xdev-developer opened 7 years ago

xdev-developer commented 7 years ago

Hi all. Unique index not working when run vertex creation in parallel.

import java.io.File
import java.util.concurrent.Executors

import org.apache.commons.configuration.BaseConfiguration
import org.apache.commons.io.{FileUtils, IOUtils}
import org.apache.tinkerpop.gremlin.structure.{Graph, Vertex}
import org.janusgraph.core.{JanusGraph, JanusGraphFactory}

import scala.concurrent.duration._
import scala.concurrent.ExecutionContext.Implicits.global
import scala.concurrent.{Await, ExecutionContext, Future}
import scala.util.Random

object Launcher {

  val DB_DIR = "/tmp/titan-index-test"

  def openGraph(): Graph = {
    val conf = new BaseConfiguration()
    conf.setProperty("schema.default", "none")
    conf.setProperty("storage.backend", "berkeleyje")
    conf.setProperty("storage.directory", DB_DIR)
    createSchema(JanusGraphFactory.open(conf))
  }

  def createSchema(graph: JanusGraph): JanusGraph = {
    val mgmt = graph.openManagement()
    val label = mgmt.makeVertexLabel("user").make()
    val key = mgmt.makePropertyKey("username").dataType(classOf[String]).make()
    val pass = mgmt.makePropertyKey("pass").dataType(classOf[String]).make()

    mgmt.buildIndex("usernameUnique", classOf[Vertex]).addKey(key).indexOnly(label).unique().buildCompositeIndex()

    mgmt.commit()
    graph
  }

  def main(args: Array[String]): Unit = {
    System.out.println("Start application")
    FileUtils.deleteQuietly(new File(DB_DIR))

    val graph = openGraph()

    //Create vertex with unique data.
    val v = graph.addVertex("user")
    v.property("pass", "test")
    v.property("username", "Test")
    graph.tx().commit()

    (1 to 1000).foreach { i =>
      println("test: " + i)

      val fSeq = Future.sequence(test(graph))

      println("Awaiting")
      Await.result(fSeq, 1.minute)

      val count = graph.traversal().V().has("username", "Den").count().next()
      println("Count of vertices: " + count)
      assert(count == 1, "Users > 1")
    }

    println("Closing application")
    graph.close()
    System.exit(0)
  }

  def test(graph: Graph)(implicit ec: ExecutionContext): List[Future[Boolean]] = {
    (1 to 10).toList.map { _ =>
      Future {
        try {
          println("Add vertex!!!")
          val v = graph.addVertex("user")
          v.property("pass", "test")
          v.property("username", "Den")
          graph.tx().commit()
          true
        } catch {
          case ex: Exception =>
            println(ex.getMessage)
            false
        }
      }
    }
  }
}

Result:

Adding this property for key [username] and value [Den] violates a uniqueness constraint [usernameUnique]
Adding this property for key [username] and value [Den] violates a uniqueness constraint [usernameUnique]
Adding this property for key [username] and value [Den] violates a uniqueness constraint [usernameUnique]
Adding this property for key [username] and value [Den] violates a uniqueness constraint [usernameUnique]
Exception in thread "main" java.lang.AssertionError: assertion failed: Users > 1
    at scala.Predef$.assert(Predef.scala:170)
    at com.xdev.titan.Launcher$$anonfun$main$1.apply$mcVI$sp(Launcher.scala:62)
    at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160)
    at com.xdev.titan.Launcher$.main(Launcher.scala:52)
    at com.xdev.titan.Launcher.main(Launcher.scala)
Count of vertices: 2

This additional vertex cannot be removed, operation failed with status NOT_FOUND.

xdev-developer commented 7 years ago

related to https://github.com/JanusGraph/janusgraph/issues/125

amcp commented 7 years ago

Do you mean related to a different issue? This issue is #121

xdev-developer commented 7 years ago

Oh, sorry wrong issue number. Comment edited. After debugging of testConcurrentConsistencyEnforcement, I can say, this is same problem.

therealnb commented 7 years ago

For reference, this happened in our deployed code also.

robopeter commented 6 years ago

Do we think this issue is specific to BerkeleyDB or might it also exist with Cassandra?

Miroka96 commented 6 years ago

It exists with Cassandra too. I have 3 times more vertices than I have unique data. Before I have created a unique index and added LOCK consistency to it. But on Spark, 96 Threads did not care about it.

ramuvistara commented 6 years ago

Hi Yeah, We are facing too same issue when vertex creation is running parallel in cassandra. Please let us know, why unique index is not working in parallel case. Is this issue is fixed?

Manson3m3 commented 6 years ago

This problem also happened when i use Hbase as backend storage.

mikehowey1 commented 5 years ago

same issue any ideas or updates??

dawany commented 5 years ago

i have the same problem when i use spark to insert vertex .Hbase is the backend storage.

To-om commented 5 years ago

Has anybody found a workaround ? I don't think that checking the existence of the vertex before creation is a valid solution.

mad commented 5 years ago

Berkeleydb just not implement locking mechanism but enable locking feature. And berkeleydb has not test suites for LockKeyColumnValueStoreTest

JG has ExpectedValueCheckingStore that's implement locking for store without locking support. Default locking implementation is redundant for local stores

  • Locking is done in two stages: first between threads inside a shared process,
  • and then between processes in a JanusGraph cluster.

Some example of only inter-process locking mechanism implemented here

priyanka211 commented 3 years ago

Hi I am also facing the same issue where unique index constraint is not working. And result is multiple vertices are getting created. But the index contains only one vertex. when I try to delete the vertices which are not in index it throws follwoijng exception: Caused by: org.janusgraph.diskstorage.locking.PermanentLockingException: Expected value mismatch for KeyColumn [k=0x 76- 15-191-228- 29-137-160- 79-114- 97- 99-108-101- 95- 66- 68- 77- 58- 47- 47- 81- 65- 49- 50- 67- 82- 49- 47- 83- 71- 83- 65- 77- 80- 76- 69- 68- 66- 47-114-101-102-116- 97-110-118-105- 47- 69- 77- 80- 78- 65- 77-197, c=0x 0]: expected=[ 24- 30- 57-200] vs actual=[ 17- 35- 28-152] (store=graphindex) at org.janusgraph.diskstorage.locking.consistentkey.ExpectedValueCheckingTransaction.checkSingleExpectedValueUnsafe(ExpectedValueCheckingTransaction.java:289) at org.janusgraph.diskstorage.locking.consistentkey.ExpectedValueCheckingTransaction.access$000(ExpectedValueCheckingTransaction.java:50) at org.janusgraph.diskstorage.locking.consistentkey.ExpectedValueCheckingTransaction$1.call(ExpectedValueCheckingTransaction.java:227) at org.janusgraph.diskstorage.locking.consistentkey.ExpectedValueCheckingTransaction$1.call(ExpectedValueCheckingTransaction.java:224) at org.janusgraph.diskstorage.util.BackendOperation.executeDirect(BackendOperation.java:69) at org.janusgraph.diskstorage.locking.consistentkey.ExpectedValueCheckingTransaction.checkSingleExpectedValue(ExpectedValueCheckingTransaction.java:224) at org.janusgraph.diskstorage.locking.consistentkey.ExpectedValueCheckingTransaction.checkAllExpectedValues(ExpectedValueCheckingTransaction.java:193) at org.janusgraph.diskstorage.locking.consistentkey.ExpectedValueCheckingTransaction.prepareForMutations(ExpectedValueCheckingTransaction.java:158) at org.janusgraph.diskstorage.locking.consistentkey.ExpectedValueCheckingStoreManager.mutateMany(ExpectedValueCheckingStoreManager.java:72) Is there a way to delete the vertices which are not there in index?

li-boxuan commented 3 years ago

Talking about Cassandra here (not familiar with BerkeleyDB):

Concurrent vertex creation should not be an issue if all competing transactions are on the same JVM due to the inter-thread locking mechanism. Concurrent vertex creation can lead to inconsistencies if competing transactions are on different JVM instances because the inter-process locking heavily relies on the data storage itself. Competing transactions basically try writing an entry to Cassandra to claim the lock. Due to the eventual consistency characteristics of Cassandra itself, it might be the case (just being hypothetical here) that more than one transaction thinks itself has got the lock, and then it goes ahead with the vertex creation, leading to data duplicates.

Note that the document has suggested that

The locking implementation is not robust against all failure scenarios. For instance, when a Cassandra cluster drops below quorum, consistency is no longer ensured. Hence, it is suggested to use locking-based consistency constraints sparingly with eventually consistent storage backends.

li-boxuan commented 3 years ago

I am also facing the same issue where unique index constraint is not working. And result is multiple vertices are getting created. But the index contains only one vertex

Hi @priyanka211 Can you give some sample code and context of how you create the vertices? Is it the case that you have multiple vertices with the same property value (which should be unique), but an index query returns only one vertex?

egetman commented 3 years ago

Hi all. The same issue =( Console output for info:

gremlin> g.V().has('Skill', 'skill_title', 'foo').elementMap()
==>{id=32808, label=Skill, skill_unique_names=foo, skill_id=e2d7a4f4-436a-44a1-b1c3-51fc5ed2037e, skill_title=foo}
gremlin> g.V().has('Skill', 'skill_title', 'Foo').elementMap()
==>{id=28712, label=Skill, skill_unique_names=foo, skill_id=14d25545-9d3e-4e8a-bd5a-2f689c4c48ea, skill_title=Foo}
gremlin> g.V().has('Skill', 'skill_unique_names', 'foo').elementMap()
==>{id=28712, label=Skill, skill_unique_names=foo, skill_id=14d25545-9d3e-4e8a-bd5a-2f689c4c48ea, skill_title=Foo}

graph.openManagement().printSchema()
...
Vertex Index Name              | Type        | Unique    | Backing        | Key:           Status |
---------------------------------------------------------------------------------------------------
...
skillByTitleExact         | Composite   | true      | internalindex  | skill_title:           ENABLED |
skillByUniqueNamesExact   | Composite   | true      | internalindex  | skill_unique_names:    ENABLED | <-- set cardinality

I reproduce the bug with topologies:

Workflow is quite simple: rapid web requests -(n requsts)-> remote client -(n requests)-> Janus

The requests are unrelated, so I can't use some simple batching (maybe if all the requests were batched in one transaction there won't be such an issue).

Sample code, how the vertex created:

public Vertex save(@Nonnull Context context) {
        final String label = context.property(LABEL_KEY);
        final Tuple<String, ?> id = requireNonNull(context.property(ID_KEY));
        try (GraphTraversal<?, Vertex> traversal = traversalSource.addV(label).property(id.getKey(), id.getValue())) {
            final Map<String, Object> vertexProperties = context.vertexProperties();
            vertexProperties.forEach((key, value) -> setProperty(traversal, key, value));
            return traversal.next();
        }
}

setProperty just perform some simple validation/cardinality logic.

li-boxuan commented 3 years ago

@egetman So you are able to reproduce the problem with single JanusGraph instance and single Cassandra? That sounds abnormal. Did you enable locking?

egetman commented 3 years ago

@li-boxuan no, I didn't.

li-boxuan commented 3 years ago

@egetman That's the problem - you should enable locking. See https://docs.janusgraph.org/advanced-topics/eventual-consistency/#data-consistency

egetman commented 3 years ago

@li-boxuan Thanks! =) It was my bad - misconfiguration...

gremlin> mgmt = graph.openManagement()
==>org.janusgraph.graphdb.database.management.ManagementSystem@7fec1b9d
gremlin> skillByUN = mgmt.getGraphIndex("skillByUniqueNamesExact")
==>skillByUniqueNamesExact
gremlin> mgmt.setConsistency(skillByUN, ConsistencyModifier.LOCK)
==>null
gremlin> mgmt.commit()
==>null

This update resolves the issue.

bishnuagrawal-zeotap commented 2 years ago

How does it work for Cassandra type of dbs? What lock mechanism is used to achieve this?