Hi,
I am checking how many "Read Count" are made per minute as follows way:
$ ./nodetool -h 127.0.0.1 cfstats
Keyspace: KSTEST
Read Count: 412303
(...)
sleep 60
$ ./nodetool -h 127.0.0.1 cfstats
Keyspace: KSTEST
Read Count: 462555
(...)
$ echo "462555 - 412303"| bc
50252
And the
> "Eric" == Eric Czech writes:
Eric> Yea that's not a mapping I'd like to maintain either -- as an
Eric> experiment, I copied production sstables to the analysis
Eric> cluster and ran brisk/cassandra without specifying an initial
Eric> token (after deleting the LocationInfo* f
I watched the logs pretty carefully as scrub ran and it definitely
skipped/deleted some problematic rows so perhaps that would explain the tmp
sstables but I'm not sure. Either way, they weren't still being written to
and after casting my restart spell, everything is totally fine and the tmp
table
tmp files are temporary in nature, if they fail to live up to their name then
they are against nature.
If they are still been written to then something is happening, either a flush
or a compaction. Check nodetool compactstats
They may indicate errors during the scrub (or compact), check the lo
What version are you on ?
The error stack is from nodetool talking to the server. Check the logs on node
3 in DC2 for errors, it sounds like perhaps it to repair or did not complete.
You can monitor a repair by looking at:
- nodetool compactionstats for a validation compaction
- nodetool netsta
It can be dangerous if wielded like a "sledge-O-matic" and not as
funny as watching Gallager smash fruit.
danger aside, I do find annotations useful for reducing lots of boiler
plate code.
On Sun, Oct 2, 2011 at 6:58 PM, aaron morton wrote:
> One thing to be careful of is giving developers the p
One thing to be careful of is giving developers the power to create CF, or a
secondary index, by creating a new Class or annotating a property with
@indexed. Adding mucho CF's and indexes can/will result in decreased
performance.
Cheers
-
Aaron Morton
Freelance Cassandra Deve
Deleting the data may not be the right approach here if you want to have a
clean slate to start the next test. It will leave tombstones around, which may
reduce your performance if you make a lot of deletes. It's pedantic, but it's
different to truncate or drop.
Truncate is doing a few more th
Yea that's not a mapping I'd like to maintain either -- as an experiment, I
copied production sstables to the analysis cluster and ran brisk/cassandra
without specifying an initial token (after deleting the LocationInfo* files
and renaming the cluster). As far as I can tell, everything is running
If you have different work loads for your data the first approach is to use
different CF's. You can make changes to the row / key cache to allocate more
cache to the important things. The OS will take care of using the rest of the
memory to cache the disk access.
In this case there is not much
Another thing Edward if you don't mind, how does cassandra choose a node to
associate with a token if there is more than one node with the same token?
I know that's definitely not a favorable situation to be in, but I'm
curious how my production ring chose to switch ownership of the tokens.
On Su
> "Eric" == Eric Czech writes:
Eric> Hi Shyamal, I was using the same cluster name but since
Eric> writing that first email, I've already had success bringing up
Eric> nodes in the analysis cluster with a different cluster name
Eric> after deleting the LocationInfo* tables.
Hi Shyamal,
I was using the same cluster name but since writing that first email, I've
already had success bringing up nodes in the analysis cluster with a
different cluster name after deleting the LocationInfo* tables.
How have you been setting the tokens in the copied version of the cluster?
A
> "Eric" == Eric Czech writes:
Eric> We're exploring a data processing procedure where we snapshot
Eric> our production cluster data and move that data to a new
Eric> cluster for analysis but I'm having some strange issues where
Eric> the analysis cluster is still somehow awar
Thanks Edward. So would you say this is a good strategy:
1. snapshot files from production cluster
2. move snapshot files to analysis cluster in a one-to-one node fashion
(the system/LocationInfo* sstables could be excluded here but I'm moving
them all because the transfer is also part of our D
On Sun, Oct 2, 2011 at 4:25 PM, Eric Czech wrote:
> We're exploring a data processing procedure where we snapshot our
> production cluster data and move that data to a new cluster for analysis but
> I'm having some strange issues where the analysis cluster is still somehow
> aware of the producti
We're exploring a data processing procedure where we snapshot our production
cluster data and move that data to a new cluster for analysis but I'm having
some strange issues where the analysis cluster is still somehow aware of the
production cluster (i.e. the production cluster ring is trying to in
Any ideas?
Thanks,
Mike Peters
On 10/1/2011 1:19 AM, Mike Peters wrote:
Hi,
We're using Cassandra 0.8 counters in production and loving it!
One issue we're running into is we need an efficient mechanism to
retrieve the "top 100" results, sorted by count values.
We have tens of thousands
If you turn the server logging up to DEBUG you will see what CL the client is
sending.
Cheers
-
Aaron Morton
Freelance Cassandra Developer
@aaronmorton
http://www.thelastpickle.com
On 1/10/2011, at 3:43 AM, Ikeda Anthony wrote:
> We are using 0.8.6
>
>
>
> On 29/09/2011, a
That is not a valid statement, were you getting an error ?
This creates the ProductCategory CF .
create column family ProductCategory
with column_type = 'Super'
and comparator = UTF8Type;
This creates the ProductCatID CF
create column family ProductCatId
WITH comparator = UTF8T
TimedOutErrors happen when less than CL nodes respond within rpc_timeout to the
coordinator. I would look at the nodetool tpstats to see if the read or write
thread pool of flooded, you will probably see a high pending count.
One way to overwhelm a node is to use a high number of record in a
21 matches
Mail list logo