1 Answer
0
I've tested a crawler using the same folder structure in S3 as mentioned.
Specified include path as: s3://my-datalake/projects/
Exclude pattern as: incremental_**/**
Using above exclude pattern ignores all files under folders named 'incremental_'. The only additional thing could be that existing crawlers have "UpdateBehavior" as "LOG" - so the already created tables are not being dropped. You could try updating it to "UPDATE_IN_DATABASE" - this will recreate the tables.
Reference - https://docs.aws.amazon.com/glue/latest/dg/define-crawler.html#crawler-data-stores-exclude
answered 4 months ago
Relevant content
- asked 12 days ago
- Accepted Answerasked 7 months ago
- Accepted Answerasked 7 months ago
- AWS OFFICIALUpdated a month ago
- AWS OFFICIALUpdated a month ago
- AWS OFFICIALUpdated 2 years ago
- AWS OFFICIALUpdated 2 years ago