-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
File not transferred from Azure to AWS when specifying a keyName for S3 not matching the part name on Azure side #238
Comments
Thanks for your contribution 🔥 We will take a look asap 🚀 |
@ank19 As per current implementation, This is the buggy code which doesn't take destination file name Lines 94 to 99 in 8fca2e5
Azure Transfer works fine as it takes destination file name |
While I haven't delved into the issue yet, there are several aspects to consider: The |
@ank19 Upon replicating the scenario in a local environment, the transfer process testing exhibited no issues when both Regarding the
|
@yurimssilva - Thanks for the update. I also think the "Invalid Upload-ID" issue from 0.6.0-rc2 is resolved. I guess that was a unlucky coincidence together with a differing behavior of the Azure data plane and the AWS data plane.
|
Hi @ank19, +1 on pushing the discussion. Would you like to follow up with that? Regarding the present issue, do you believe it is still beneficial to keep it open or can it be closed? |
Hi @rafaelmag110 - I'll suggest that you start that discussion from your side, mainly for the reason that we on our side are not using AWS. I'm closing that issue, but I think it would be benefitial though, to align the behavior for both data planes to avoid future misunderstanding. |
Bug Report
Describe the Bug
Expected Behavior
A consumer willing to transfer a file to it's own AWS S3 bucket from a provider, which is using Azure Data Lake as a source, should be able to specify a "keyName" in the S3 data transfer destination, which does not necessarily have to exactly match the object name from the asset specification used on provider side within the Azure Data Lake.
Observed Behavior
Assuming the provider created an asset on Azure Data Lake in a directory "Outbound/4MB" and assuming that the consumer specifies keyName="Outbound/4MB", then we're able to transfer the file successfully.
If the consumer however changes the keyName to something not exactly matching that, e. g. by removing the directory name, changing the file name, then the transfer process does not succeed (but marked as COMPLETED).
Steps to Reproduce
Steps to reproduce the behavior:
"dataAddress": {
"@type": "DataAddress",
"type": "AzureStorage",
"container": "...",
"account": "...",
"blobName": "Outbound/4MB.bin"
}
Context Information
Detailed Description
Further notes:
Possible Implementation
I had a look at the implementation, but I haven't found a hint yet. The S3DataSink does not seem to use the keyName at all (or just for logging purposes before @yurimssilva fix for the exception handling). The S3DataSinkFactory only seems to use the keyName for accessing the vault, which is not relevant in our case.
The text was updated successfully, but these errors were encountered: