Troubleshooting - Neptune Analytics

Troubleshooting

For both bulk load and batch load, all the errors and summary of the load is sent to the CloudWatch log group in your account. To view the logs go to CloudWatch, click log groups from the left column, then search for and click /aws/neptune/import-task-logs/.

  1. Batch Load: The logs for each load is saved under /aws/neptune/import-task-logs/<graph-id>/<load-id> CloudWatch log stream.

  2. Bulk Load using Import Task: The logs are saved under /aws/neptune/import-task-logs/<graph-id>/<task-id> CloudWatch log stream.

  • S3_ACCESS_DENIED: The server does not have permissions to list or download the given file. Fix the permissions and retry. See Create your IAM role for Amazon S3 access for help setting up the Amazon S3 permissions.

  • LARGE_STRING_ERROR: One or more strings exceeded the limit on the size of strings. This data cannot be inserted as is. Update the strings exceeding the limit and retry.

  • PARSING_ERROR: Error parsing the given value(s). Correct the value(s) and retry. More information on different parsing errors is provided in this section.

  • OUT_OF_MEMORY: No more data can be loaded in the current m-NCU. If encountered during import task, set a higher m-NCU and retry. If encountered during batch load, scale the number of m-NCU and retry the batch load.

  • PARTITION_FULL_ERROR: No more data can be loaded in the internal server configuration. If encountered during import task, the import workflow would change the server configuration and retry. If encountered during batch load, reach out to the AWS service team to unblock loading of new data.

Common parsing errors and solutions

Error template Solution

Invalid data type encountered for header val:badtype when parsing line [:ID,firstName:String,val:badtype,:LABEL].

Incorrect Datatype provided. Check the documentation for supported data types. See Data formats for more information.

Multi-valued columns are not supported firstName:String[] when parsing line [:ID,firstName:String[],val:String,:LABEL].

The opencypher format does not support multivalued user defined properties. Try using the csv format to insert multivalued vertex properties, or remove multivalued properties.

Bad header for a file in 'OPEN_CYPHER' format, could not determine node or relationship file, found system columns from 'csv' format when parsing line [~id,firstName:String,val:int,:LABEL].

Both the opencypher and csv format expect certain header columns to be present. Make sure you have entered them correctly.

Check the Data formats documentation for required fields by format.

Bad header for a file in 'OPEN_CYPHER' format, could not determine node or relationship file.

The header of the files does not have the required system columns. Check the Data formats for required fields by format.

Relationship file in 'OPEN_CYPHER' format should contain both :START_ID and :END_ID columns when parsing line [:START_ID,firstName:String].

The header of the edge files does not have all the required system columns. Check the Data formats for required fields by format.

Invalid data type. Found system columns from 'OPEN_CYPHER' format :ID when parsing line [:ID,firstName:String,val:Int,~label].

The opencypher and csv formats have different system column names, and they begin with : and ~ respectively. User defined properties cannot begin with those reserved prefixes in the respective formats. Confirm the format name and system column names, or update user defined properties to not use reserved prefixes.

Named column name is not present for header field :BLAH when parsing line [:ID,:BLAH,firstName:String].

The opencypher and csv formats have different system column names, and they begin with : and ~ respectively. User defined properties cannot begin with those reserved prefixes in the respective formats. Confirm the format name and system column names, or update user defined properties to not use reserved prefixes.

System column other than ID cannot be stored as a property: <columnHeader>.

The opencypher and csv formats have different system column names, and they begin with : and ~ respectively. User defined properties cannot begin with those reserved prefixes in the respective formats. Confirm the format name and system column names, or update user defined properties to not use reserved prefixes.

Duplicate user column firstName when parsing line [:ID,:LABEL, firstName:String, firstName:String].

The file contains duplicate user defined property column names in the header. Remove all of the duplicate columns.

Duplicate system column :ID found when parsing line [:ID,:ID,firstName:String,:LABEL].

The file contains duplicate system column names in the header. Remove all of the duplicate columns.

Invalid column name provided for loading embeddings: [abcd] for filename: someFilename. Embedding column name must be the same as their corresponding vector index name when parsing line [:ID,firstName:String,abcd:Vector,:LABEL] in [filename].

An incorrect name is used for the vector embeddings.

"date" type is curretly not supported. "datetime" may be an alternative type.

Use datetime as the field type as date type suppoorted yet in Neptune Analytics.

Headers must be non-empty.

Headers need to be non empty. If the file has an empty line in the beginning, remove the empty line.

Failure encounted while parsing the csv file.

Likely reason is the number of columns in the row doesn't match the number of columns provided in the header. If you dont have a value for a column, provide an empty value.

For example: 123,vertex,,,.

Could not process value of type:http://www.w3.org/2001/XMLSchema#int for value: a when parsing line [v1,v19683,con,a] in [file].

There is a mismatch between the type of the value provided for that column in the row and the type specified in the header. In this specific case the column header is annotated with integer type but a is not parseable as an integer.

Could not load vector embedding: [a,bc]. Check the dimensionality for this vector.

The size of the vector does not match the dimension defined in the vector search configuration for the graph.

Could not load vector embedding: [a,NaN]. Check the value for this vector.

Float and double values in scientific notation are currently not supported. Also Infinity, -Infinity, INF, -INF, and NaN are not recognized.