[ 
https://issues.apache.org/jira/browse/CASSANDRA-21209?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=18065358#comment-18065358
 ] 

Stefan Miklosovic commented on CASSANDRA-21209:
-----------------------------------------------

Hi [~jeffzhong] thanks for your help but I have already PR up if I am not 
mistaken? We also need to refactor it more deeply and I am working on it as we 
speak. 

> Fix exception when training ZSTD compression dictionary and using bigger 
> buffer for sampling data than possible to create
> -------------------------------------------------------------------------------------------------------------------------
>
>                 Key: CASSANDRA-21209
>                 URL: https://issues.apache.org/jira/browse/CASSANDRA-21209
>             Project: Apache Cassandra
>          Issue Type: Bug
>          Components: Feature/Compression
>            Reporter: Stefan Miklosovic
>            Assignee: Jeff Zhong
>            Priority: Normal
>             Fix For: 5.x
>
>          Time Spent: 20m
>  Remaining Estimate: 0h
>
> We can specify how much sampling data there should be for training of a 
> dictionary. The buffer for training is a direct buffer. If we say that we 
> will be training on 2GiB, then it will try to create a direct buffer of size 
> 2GiB.
> This problem is visible e.g. when starting Cassandra via IDEA which uses 1GiB 
> heap. I think how it works is that max direct memory size for JVM is 
> basically same as xmx so it will fail. 
> The fix would consist of wrapping creation of that buffer in a try-catch and 
> propagate the error in some sanitized way up to a caller informing them it is 
> not possible to create such a big sampling buffer.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to