Can you please post the result of the following code?val list = new Neo4jTable(StructType(Array(StructField("foo", DataTypes.StringType))), Map("url" -> "bolt://neo4j:7687", "labels" -> "foo").asJava, "").capabilities()println(list)
@brianmartin the best practice on batch importing the data with Spark is:insert in parallel the nodes by partitioning the data via the node key column (otherwise this will lead to locking issues and you cannot leverage the parallelism); please consid...
There are a ton of reasons that can contribute to slow down the process:Neo4j hardware issues:is the HD fast enough? is the RAM enough?If you reuse the same Spark DataFrame over the time and you don't cache it this forces Spark to recompute it each t...