Skip to content

Add big dumper (delta lake table export)#220

Draft
bdpedigo wants to merge 8 commits intomasterfrom
delta-out
Draft

Add big dumper (delta lake table export)#220
bdpedigo wants to merge 8 commits intomasterfrom
delta-out

Conversation

@bdpedigo
Copy link
Copy Markdown

No description provided.



@celery.task(
name="deltalake:write_deltalake_table",
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

you should use an existing queue so that it gets run on an existing set of workers... i think you want orchestration for long running jobs

Copy link
Copy Markdown
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

will this allow us to still configure the workers (e.g. amount of memory) or will that make them tied to whatever other precesses are in this queue?

sql_uri_config, datastack, version
)

output_bucket = get_config_param("DELTALAKE_OUTPUT_BUCKET")
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this should be just a different path off of the csv dump

Copy link
Copy Markdown
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

does said path need to be configurable or can it be hard coded

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants