We think it is this bug:
https://issues.apache.org/jira/browse/CASSANDRA-8860
We're rolling a patch to beta before rolling it into production.
On Wed, Mar 4, 2015 at 4:12 PM, graham sanderson wrote:
> We can confirm a problem on 2.1.3 (sadly our beta sstable state obviously
> did not match our
We can confirm a problem on 2.1.3 (sadly our beta sstable state obviously did
not match our production ones in some critical way)
We have about 20k sstables on each of 6 nodes right now; actually a quick
glance shows 15k of those are from OpsCenter, which may have something to do
with beta/prod
Are you finding a correlation between the shards on the OOM DC1 nodes and
the OOM DC2 nodes? Does your monitoring tool indicate that the DC1 nodes
are using significantly more CPU (and memory) than the nodes that are NOT
failing? I am leading you down the path to suspect that your sharding is
givin
What kind of disks are you running here? Are you getting a lot of GC before
the OOM?
Patrick
On Wed, Mar 4, 2015 at 9:26 AM, Jan wrote:
> HI Roni;
>
> You mentioned:
> DC1 servers have 32GB of RAM and 10GB of HEAP. DC2 machines have 16GB of
> RAM and 5GB HEAP.
>
> Best practices would be be to:
HI Roni;
You mentioned: DC1 servers have 32GB of RAM and 10GB of HEAP. DC2 machines have
16GB of RAM and 5GB HEAP.
Best practices would be be to:a) have a consistent type of node across both
DC's. (CPUs, Memory, Heap & Disk)
b) increase heap on DC2 servers to be 8GB for C* Heap
The leveled
Hi there,
We are running C* 2.1.3 cluster with 2 DataCenters: DC1: 30 Servers /
DC2 - 10 Servers.
DC1 servers have 32GB of RAM and 10GB of HEAP. DC2 machines have 16GB
of RAM and 5GB HEAP.
DC1 nodes have about 1.4TB of data and DC2 nodes 2.3TB.
DC2 is used only for backup purposes. There are no re