All Content tagged with AWS Glue DataBrew
AWS Glue DataBrew is a new visual data preparation tool that makes it easy for data analysts and data scientists to clean and normalize data to prepare it for analytics and machine learning.
Content language: English
Select up to 5 tags to filter
Sort by most recent
Why don't I see the column - ruleset_name, in the output of s3 file generated by a Glue Data Quality job ? I see the below columns in the JSON output of the Glue DQ job. Is there any way, I can get...
Hello All ,
I am trying to clean up my dataset see below ![Dataset](/media/postImages/original/IMlISUqk8QRVCZTKGDM3uwcA).
I want to remove the first row since the name is invalid and want to add it...
I am closely following the Data Analysis and Visualization in AWS wokrshop. Once I create a job in Glue Databrew and select the role that we set up with the permission given by the workshop, I get...
Hi team,
I am running the data quality rules over my dataset in databrew and getting the dq results in JSON format which consist with the pointers of the schema information about my data ( column...
Hello, I would like to know if there is a way to query Iceberg tables (backed with S3 parquet files) cataloged within the AWS Glue Catalog using AWS Databrew. (maybe through Athena?).
Also, is it...
Hi,
I'm trying to programmatically kick off a DataBrew profile job using AWS SDK from my java application. I need to profile MySQL database tables. While I'm able to do that from my application, I...
Dear AWS Support Team,
I am are currently implementing a data governance tool utilizing AWS Lake Formation and AWS Glue Databrew for data transformations. i've encountered an issue: Glue Databrew...
Hi,
When i open an already existing project, it always crash at the recipe validation step. Therefore i cannot do any modifications on the recipe when i open my project. Because the recipe is not...
![Enter image description here](/media/postImages/original/IM6awzeeuxT32zW98DS9S78Q)
Hello, I have recently had to move one of our Jobs from Glue DataBrew to Glue ETL - which is now up and running.
The issue I am having is that everytime the job runs it creates a new series of files....
Hello, we are receiving the below error on one of our jobs:
RecipeStepError: Failed at step 2. Parameters: {'operation': 'JSON_TO_STRUCTS', 'sourceColumns': '["data"]', 'unnestLevel': '120'}. Error:...
We have some business logic we would like to implement in DataBrew, that would be difficult to do using the existing predefined operations. Is it possible to add a custom build operation to do this...