- Newest
- Most votes
- Most comments
I apologize, but there seems to be a misunderstanding in your approach. The Neo4j Connector for Apache Spark is not directly compatible with Amazon Neptune for loading data. While Neptune does provide some compatibility with Neo4j tools, the Spark connector is not one of them.
For loading data into Neptune using Spark, you should use Neptune's own integration methods. Neptune integrates with Apache Spark through an open-source library called neptune-python-utilities, which can be used for running Python and Gremlin jobs.
The error you're encountering ("There is no procedure with the name dbms.components
registered") is likely because Neptune doesn't support Neo4j-specific procedures and features.
To load data into Neptune using Spark, you would typically use Neptune's bulk loader API or other AWS services like AWS Glue. Here's a general approach you could consider:
- Prepare your data in Spark as a DataFrame.
- Use AWS Glue or custom ETL processes to transform your data into a format compatible with Neptune (CSV for bulk loading or Gremlin/SPARQL for individual inserts).
- Use Neptune's bulk loader API or write custom code to insert the data using Gremlin or SPARQL queries.
If you need to use Spark specifically, you might want to look into using the AWS Glue Neptune connector, which provides a way to read from and write to Neptune using Spark jobs in AWS Glue.
For the most up-to-date and accurate information on loading data into Neptune, I recommend checking the official AWS documentation or contacting AWS support for guidance tailored to your specific use case.
Sources
Neptune compatibility with Neo4j - Amazon Neptune
Working with other AWS services - Amazon Neptune
I think we might want to remove that it is compatible in the documentation in that case.
Relevant content
- asked 3 months ago
- asked a year ago
It looks like you are using the 5.X driver series. Can you try with the 4.X driver series as shown in the docs and see if that solves this issue?
<dependency> <groupId>org.neo4j</groupId> <artifactId>neo4j-connector-apache-spark_2.12-4.1.0</artifactId> <version>4.0.1_for_spark_3</version> </dependency>