Query editor v2 on Redshift show Error:Parsed manifest is not a valid JSON object

0

i'm used AWS glue crawler data from S3 Bucket to Datacata log and using AWS redshift query but show error: ERROR: Parsed manifest is not a valid JSON object. Detail: ----------------------------------------------- error: Parsed manifest is not a valid JSON object. code: 15003 context: staff_id,first_name,last_name,email,phone,active,store_id,manager_id 1,Fabiola,Jackson,fabiola.jackson@bikes.shop,(831) 555-5554,1,1,NULL 2,Mireya,Copeland,mireya.copeland@bikes.shop,(831) 555-55 query: 209932[child_sequence:1] location: spectrum_manifest.cpp:73 process: padbmaster [pid=1073873287] ----------------------------------------------- [ErrorId: 1-6569e828-0a2aeb393ad3283c4d1eea4f] Please help me check thank you very much

asked 5 months ago355 views
1 Answer
3

Hello,

This issue usually happens for a couple of reasons.

  1. Table associated with the database might not have the permissions to allow the specified user, querying the table, to perform the necessary Lake Formation actions. Granting the user who executes the query, to needed Lake Formation permissions on the given table for the database might fix the issue.
  2. Table that had been created with their LOCATION set to a specific file in an S3 path (rather than an S3 path prefix) might fail with above error. Based on this I would recommend placing the files for each table into separate subfolders(instead of specific file) to make the data queryable as described in this article

To recrawl the s3 data source with specific s3 prefix as shown below,

  1. Navigate to the Glue Crawler Page.

  2. Click on Create Crawler, and enter a crawler name and click Next.

  3. Click on Add a Data Source.

  4. Choose S3 as the data source and the include path of your files 's3://<bucket-name>/data_table/' (Please note that folder 'data_table' contains source files) and click on Update S3 data source. Click Next

  5. Create an IAM Role and enter a name and click Next.

  6. Click on Add database and enter a database name.

  7. Select Run on demand for the frequency and click Next.

  8. Select all remaining defaults. Once the Crawler has been created, click on Run Crawler.

  9. Once the Crawler has completed its run, you will see a new table in the Glue Catalo

  10. Run 'Select' command in Redshift

Reference here

AWS
SUPPORT ENGINEER
answered 5 months ago

You are not logged in. Log in to post an answer.

A good answer clearly answers the question and provides constructive feedback and encourages professional growth in the question asker.

Guidelines for Answering Questions