[
https://issues.apache.org/jira/browse/HADOOP-14660?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16109489#comment-16109489
]
Steve Loughran commented on HADOOP-14660:
-----------------------------------------
I'm getting a test failure here, with trunk + patch 010; not (yet) tested with
trunk without the patch.
{code}
-------------------------------------------------------
Running org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper
Tests run: 10, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 149.902 sec
<<< FAILURE! - in org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper
testWhenOneInstanceIsDown(org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper)
Time elapsed: 4.483 sec <<< FAILURE!
org.mockito.exceptions.verification.TooLittleActualInvocations:
httpClient.execute(
<Http get for service local>
);
Wanted 2 times:
-> at
org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper.testWhenOneInstanceIsDown(TestWasbRemoteCallHelper.java:359)
But was 1 time:
-> at
org.apache.hadoop.fs.azure.WasbRemoteCallHelper.retryableRequest(WasbRemoteCallHelper.java:151)
at
org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper.testWhenOneInstanceIsDown(TestWasbRemoteCallHelper.java:359)
Results :
Failed tests:
TestWasbRemoteCallHelper.testWhenOneInstanceIsDown:359
httpClient.execute(
<Http get for service local>
);
Wanted 2 times:
-> at
org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper.testWhenOneInstanceIsDown(TestWasbRemoteCallHelper.java:359)
But was 1 time:
-> at
org.apache.hadoop.fs.azure.WasbRemoteCallHelper.retryableRequest(WasbRemoteCallHelper.java:151)
Tests run: 10, Failures: 1, Errors: 0, Skipped: 0
{code}
> wasb: improve throughput by 34% when account limit exceeded
> -----------------------------------------------------------
>
> Key: HADOOP-14660
> URL: https://issues.apache.org/jira/browse/HADOOP-14660
> Project: Hadoop Common
> Issue Type: Improvement
> Components: fs/azure
> Reporter: Thomas Marquardt
> Assignee: Thomas Marquardt
> Attachments: HADOOP-14660-001.patch, HADOOP-14660-002.patch,
> HADOOP-14660-003.patch, HADOOP-14660-004.patch, HADOOP-14660-005.patch,
> HADOOP-14660-006.patch, HADOOP-14660-007.patch, HADOOP-14660-008.patch,
> HADOOP-14660-010.patch
>
>
> Big data workloads frequently exceed the Azure Storage max ingress and egress
> limits
> (https://docs.microsoft.com/en-us/azure/azure-subscription-service-limits).
> For example, the max ingress limit for a GRS account in the United States is
> currently 10 Gbps. When the limit is exceeded, the Azure Storage service
> fails a percentage of incoming requests, and this causes the client to
> initiate the retry policy. The retry policy delays requests by sleeping, but
> the sleep duration is independent of the client throughput and account limit.
> This results in low throughput, due to the high number of failed requests
> and thrashing causes by the retry policy.
> To fix this, we introduce a client-side throttle which minimizes failed
> requests and maximizes throughput. Tests have shown that this improves
> throughtput by ~34% when the storage account max ingress and/or egress limits
> are exceeded.
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]