Replies: 7 comments
-
Thanks for opening your first issue here! Be sure to follow the issue template! |
Beta Was this translation helpful? Give feedback.
-
It would be a good idea to check with Minio maintainers and potentially Amazon about this. Would you be interested in doing this? |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Update: This only happens when the connection is specified using the URI format (through environment variables). When using the json encoding everything works fine. |
Beta Was this translation helpful? Give feedback.
-
Could you type in exact example of the URL/variable you use ? |
Beta Was this translation helpful? Give feedback.
-
Hey @spelufo, As @potiuk suggested, can you provide screenshots or text snippet of exactly how you're configuring the connection for logging as well as the config necessary for remote logging (i.e. how you're specifying I tested on my end specifying @eladkal, @kaxil |
Beta Was this translation helpful? Give feedback.
-
Since there was no answer from the user when asked twice, converting it into discussion. |
Beta Was this translation helpful? Give feedback.
-
Apache Airflow version
2.3.1 (latest released)
What happened
We've been trying to configure remote_logging to minio with the s3_task_handler, as described here. We test on a local k8s kind cluster, with the airflow helm chart. After some debugging and verifying that the connection can be made and authenticated, we see the following from the minio logs:
You'll note that the url contains
s3%3A//
. I think this is the cause of the bad request. Running the following snippet from the airflow webserver pod in a python shell reproduces the bad request, while doing the same without the prefix succeeds.In this test I see that airflow might think having the prefix in the key is fine. The moto mock probably handles it, perhaps s3 does as well. Minio doesn't seem to handle it. The S3 documentation for object keys seem to imply that the prefix is not part of the key.
We've tried removing the prefix from the
remote_base_log_folder
which reading through the code looks like it is the source of the value ofkey
, but then airflow has no way of knowing what logging backend to use, and running migrations fails.What you think should happen instead
Airflow shouldn't pass s3 keys with the
s3://
prefix tohead_object
.How to reproduce
No response
Operating System
Kind cluster, kubernetes 1.20.7
Versions of Apache Airflow Providers
Deployment
Official Apache Airflow Helm Chart
Deployment details
No response
Anything else
No response
Are you willing to submit PR?
Code of Conduct
Beta Was this translation helpful? Give feedback.
All reactions