Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

freeze AWS glue job #193

Open
andreisorokin-astrazeneca opened this issue Jan 17, 2024 · 2 comments
Open

freeze AWS glue job #193

andreisorokin-astrazeneca opened this issue Jan 17, 2024 · 2 comments

Comments

@andreisorokin-astrazeneca

I use WatchTower as workaround for inability to change CloudWatch group for glue jobs written in Python.
But when I add WatchTower handler the glue jobs after successfuly finishing job freeze.

In Spark worker logs I see nothing - just normal finishing and after that final message that the worker had been killed by driver.

In Spark Driver I see repeated attempts to send stdout to CloudWatch by internal AWS mechanism - do not see how is possibly connected with WatchTower:

INFO LogPusher: uploading /tmp/spark-event-logs/ to s3://****
INFO MultipartUploadOutputStream: close closed:false s3://***.inprogress

If I set use_queues=False it finishes normally and in logs I see
INFO MultipartUploadOutputStream: close closed:false s3://***/FilteredOut/run-1705485647329-part-r-00002.gz

@andreisorokin-astrazeneca
Copy link
Author

... and in both cases logs were successfully and fully sent to CloudWatch by WatchTower

@andreisorokin-astrazeneca
Copy link
Author

andreisorokin-astrazeneca commented Jan 19, 2024

I found workaround - if I add that in the end of glue job

logger = logging.getLogger()
logging.shutdown()
for handler in logger.handlers:
    logger.removeHandler(handler)

it finishes normally

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant