[ 
https://issues.apache.org/jira/browse/HADOOP-14660?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16109489#comment-16109489
 ] 

Steve Loughran commented on HADOOP-14660:
-----------------------------------------

I'm getting a test failure here, with trunk + patch 010; not (yet) tested with 
trunk without the patch. 
{code}
-------------------------------------------------------
Running org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper
Tests run: 10, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 149.902 sec 
<<< FAILURE! - in org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper
testWhenOneInstanceIsDown(org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper)  
Time elapsed: 4.483 sec  <<< FAILURE!
org.mockito.exceptions.verification.TooLittleActualInvocations: 
httpClient.execute(
    <Http get for service local>
);
Wanted 2 times:
-> at 
org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper.testWhenOneInstanceIsDown(TestWasbRemoteCallHelper.java:359)
But was 1 time:
-> at 
org.apache.hadoop.fs.azure.WasbRemoteCallHelper.retryableRequest(WasbRemoteCallHelper.java:151)

        at 
org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper.testWhenOneInstanceIsDown(TestWasbRemoteCallHelper.java:359)


Results :

Failed tests: 
  TestWasbRemoteCallHelper.testWhenOneInstanceIsDown:359 
httpClient.execute(
    <Http get for service local>
);
Wanted 2 times:
-> at 
org.apache.hadoop.fs.azure.TestWasbRemoteCallHelper.testWhenOneInstanceIsDown(TestWasbRemoteCallHelper.java:359)
But was 1 time:
-> at 
org.apache.hadoop.fs.azure.WasbRemoteCallHelper.retryableRequest(WasbRemoteCallHelper.java:151)


Tests run: 10, Failures: 1, Errors: 0, Skipped: 0

{code}

> wasb: improve throughput by 34% when account limit exceeded
> -----------------------------------------------------------
>
>                 Key: HADOOP-14660
>                 URL: https://issues.apache.org/jira/browse/HADOOP-14660
>             Project: Hadoop Common
>          Issue Type: Improvement
>          Components: fs/azure
>            Reporter: Thomas Marquardt
>            Assignee: Thomas Marquardt
>         Attachments: HADOOP-14660-001.patch, HADOOP-14660-002.patch, 
> HADOOP-14660-003.patch, HADOOP-14660-004.patch, HADOOP-14660-005.patch, 
> HADOOP-14660-006.patch, HADOOP-14660-007.patch, HADOOP-14660-008.patch, 
> HADOOP-14660-010.patch
>
>
> Big data workloads frequently exceed the Azure Storage max ingress and egress 
> limits 
> (https://docs.microsoft.com/en-us/azure/azure-subscription-service-limits).  
> For example, the max ingress limit for a GRS account in the United States is 
> currently 10 Gbps.  When the limit is exceeded, the Azure Storage service 
> fails a percentage of incoming requests, and this causes the client to 
> initiate the retry policy.  The retry policy delays requests by sleeping, but 
> the sleep duration is independent of the client throughput and account limit. 
>  This results in low throughput, due to the high number of failed requests 
> and thrashing causes by the retry policy.
> To fix this, we introduce a client-side throttle which minimizes failed 
> requests and maximizes throughput.  Tests have shown that this improves 
> throughtput by ~34% when the storage account max ingress and/or egress limits 
> are exceeded. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to