Hello!
We are experiencing error messages from the client that we don’t totally
understand. They look like the following:
<Riak::ProtobuffsErrorResponse: Expected success from Riak but received 1013.
no response from backend>
Checking the riak error and crash logs, I’m seeing “overload” errors which I
assume is causing the “no response from backend” client errors:
{error,
badarg,
[{erlang,iolist_to_binary,[overload],[]},
{riak_kv_ts_svc,make_rpberrresp,2,[{file,"src/riak_kv_ts_svc.erl"},{line,483}]},
{riak_kv_ts_svc,sub_tsqueryreq,4,[{file,"src/riak_kv_ts_svc.erl"},{line,445}]},
{riak_kv_pb_ts,process,2,[{file,"src/riak_kv_pb_ts.erl"},{line,71}]},
{riak_api_pb_server,process_message,4,[{file,"src/riak_api_pb_server.erl"},{line,388}]},
{riak_api_pb_server,connected,2,[{file,"src/riak_api_pb_server.erl"},{line,226}]},
{riak_api_pb_server,decode_buffer,2,[{file,...},...]},...]}
I’m curious if these overload errors are caused by clients requesting more
concurrent TS queries than our current setting for
timeseries_max_concurrent_queries allows OR if the the
timeseries_max_concurrent_queries is set too high and we are causing riak to
crash.
Do you have any recommendations on what timeseries_max_concurrent_queries
should be set to relative to hardward specs? I assume it should be limited
based on disk I/O bandwidth.
Also, does anyone have any recommendations on query pooling so we can guarantee
that multiple clients will not generate more queries than the cluster can
handle? I like HAProxy for HTTP connection pooling but it doesn’t seem like it
would work well for limiting the number of global queries from multiple PBC
clients.
Thank you!
Chris
_______________________________________________
riak-users mailing list
[email protected]
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com