riak_test/tests/yz_handoff.erl
Zeeshan Lakhani 54568a0c27 * rename file to reduce conflict on buildbot runs
* wait_for_index on write data due to use in upgrade tests
* clean-up other tests to that test upgrade from (2.0.5->2.0.6/2.1.1->2.1.2), namely yz_extractors
* clean-up tests that can start using yokozuna_rt helpers
2015-05-27 10:25:05 -04:00

209 lines
7.6 KiB
Erlang

%% -------------------------------------------------------------------
%%
%% Copyright (c) 2014 Basho Technologies, Inc.
%%
%% This file is provided to you under the Apache License,
%% Version 2.0 (the "License"); you may not use this file
%% except in compliance with the License. You may obtain
%% a copy of the License at
%%
%% http://www.apache.org/licenses/LICENSE-2.0
%%
%% Unless required by applicable law or agreed to in writing,
%% software distributed under the License is distributed on an
%% "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
%% KIND, either express or implied. See the License for the
%% specific language governing permissions and limitations
%% under the License.
%%
%%-------------------------------------------------------------------
-module(yz_handoff).
-compile(export_all).
-include_lib("eunit/include/eunit.hrl").
-include_lib("riakc/include/riakc.hrl").
-define(GET(K,L), proplists:get_value(K, L)).
-define(FMT(S, Args), lists:flatten(io_lib:format(S, Args))).
-define(INDEX, <<"test_idx">>).
-define(BUCKET, <<"test_bkt">>).
-define(NUMRUNSTATES, 1).
-define(SEQMAX, 1000).
-define(TESTCYCLE, 20).
-define(N, 3).
-define(CFG,
[
{riak_core,
[
{ring_creation_size, 16},
{n_val, ?N},
{handoff_concurrency, 10},
{vnode_management_timer, 1000}
]},
{riak_kv,
[
%% allow AAE to build trees and exchange rapidly
{anti_entropy_build_limit, {100, 1000}},
{anti_entropy_concurrency, 8},
{handoff_rejected_max, infinity}
]},
{yokozuna,
[
{anti_entropy_tick, 1000},
{enabled, true}
]}
]).
-record(trial_state, {
solr_url_before,
solr_url_after,
leave_node,
join_node,
admin_node}).
confirm() ->
%% Setup cluster initially
[Node1, Node2, _Node3, _Node4, _Node5] = Nodes = rt:build_cluster(5, ?CFG),
rt:wait_for_cluster_service(Nodes, yokozuna),
ConnInfo = ?GET(Node2, rt:connection_info([Node2])),
{Host, Port} = ?GET(http, ConnInfo),
Shards = [{N, node_solr_port(N)} || N <- Nodes],
%% Generate keys, YZ only supports UTF-8 compatible keys
Keys = [<<N:64/integer>> || N <- lists:seq(1, ?SEQMAX),
not lists:any(fun(E) -> E > 127 end,
binary_to_list(<<N:64/integer>>))],
KeyCount = length(Keys),
Pid = rt:pbc(Node2),
yokozuna_rt:write_data(Nodes, Pid, ?INDEX, ?BUCKET, Keys),
timer:sleep(1100),
%% Separate out shards for multiple runs
[Shard1|Shards2Rest] = Shards,
{_, SolrPort1} = Shard1,
[{_, SolrPort2}|_] = Shards2Rest,
SolrURL = internal_solr_url(Host, SolrPort1, ?INDEX, Shards),
BucketURL = bucket_keys_url(Host, Port, ?BUCKET),
SearchURL = search_url(Host, Port, ?INDEX),
lager:info("Verify Replicas Count = (3 * docs/keys) count"),
verify_count(SolrURL, (KeyCount * ?N)),
States = [#trial_state{solr_url_before = SolrURL,
solr_url_after = internal_solr_url(Host, SolrPort2, ?INDEX, Shards2Rest),
leave_node = Node1},
#trial_state{solr_url_before = internal_solr_url(Host, SolrPort2, ?INDEX, Shards2Rest),
solr_url_after = SolrURL,
join_node = Node1,
admin_node = Node2}],
%% Run Shell Script to count/test # of replicas and leave/join
%% nodes from the cluster
[[begin
check_data(Nodes, KeyCount, BucketURL, SearchURL, State),
check_counts(Pid, KeyCount, BucketURL)
end || State <- States]
|| _ <- lists:seq(1,?NUMRUNSTATES)],
pass.
%%%===================================================================
%%% Private
%%%===================================================================
node_solr_port(Node) ->
{ok, P} = riak_core_util:safe_rpc(Node, application, get_env,
[yokozuna, solr_port]),
P.
internal_solr_url(Host, Port, Index) ->
?FMT("http://~s:~B/internal_solr/~s", [Host, Port, Index]).
internal_solr_url(Host, Port, Index, Shards) ->
Ss = [internal_solr_url(Host, ShardPort, Index)
|| {_, ShardPort} <- Shards],
?FMT("http://~s:~B/internal_solr/~s/select?wt=json&q=*:*&shards=~s",
[Host, Port, Index, string:join(Ss, ",")]).
%% @private
bucket_keys_url(Host, Port, BName) ->
?FMT("http://~s:~B/buckets/~s/keys?keys=true", [Host, Port, BName]).
%% @private
search_url(Host, Port, Index) ->
?FMT("http://~s:~B/solr/~s/select?wt=json&q=*:*", [Host, Port, Index]).
verify_count(Url, ExpectedCount) ->
AreUp =
fun() ->
{ok, "200", _, DBody} = ibrowse:send_req(Url, [], get, []),
FoundCount = get_count(DBody),
lager:info("FoundCount: ~b, ExpectedCount: ~b",
[FoundCount, ExpectedCount]),
ExpectedCount =:= FoundCount
end,
?assertEqual(ok, rt:wait_until(AreUp)),
ok.
get_count(Resp) ->
Struct = mochijson2:decode(Resp),
kvc:path([<<"response">>, <<"numFound">>], Struct).
get_keys_count(BucketURL) ->
{ok, "200", _, RBody} = ibrowse:send_req(BucketURL, [], get, []),
Struct = mochijson2:decode(RBody),
length(kvc:path([<<"keys">>], Struct)).
check_counts(Pid, InitKeyCount, BucketURL) ->
PBCounts = [begin {ok, Resp} = riakc_pb_socket:search(
Pid, ?INDEX, <<"*:*">>),
Resp#search_results.num_found
end || _ <- lists:seq(1,?TESTCYCLE)],
HTTPCounts = [begin {ok, "200", _, RBody} = ibrowse:send_req(
BucketURL, [], get, []),
Struct = mochijson2:decode(RBody),
length(kvc:path([<<"keys">>], Struct))
end || _ <- lists:seq(1,?TESTCYCLE)],
MinPBCount = lists:min(PBCounts),
MinHTTPCount = lists:min(HTTPCounts),
lager:info("Before-Node-Leave PB: ~b, After-Node-Leave PB: ~b",
[InitKeyCount, MinPBCount]),
?assertEqual(InitKeyCount, MinPBCount),
lager:info("Before-Node-Leave PB: ~b, After-Node-Leave HTTP: ~b",
[InitKeyCount, MinHTTPCount]),
?assertEqual(InitKeyCount, MinHTTPCount).
check_data(Cluster, KeyCount, BucketURL, SearchURL, S) ->
CheckCount = KeyCount * ?N,
KeysBefore = get_keys_count(BucketURL),
UpdatedCluster = leave_or_join(Cluster, S),
yokozuna_rt:wait_for_aae(UpdatedCluster),
KeysAfter = get_keys_count(BucketURL),
lager:info("KeysBefore: ~b, KeysAfter: ~b", [KeysBefore, KeysAfter]),
?assertEqual(KeysBefore, KeysAfter),
lager:info("Verify Search Docs Count =:= key count"),
lager:info("Run Search URL: ~s", [SearchURL]),
verify_count(SearchURL, KeysAfter),
lager:info("Verify Replicas Count = (3 * docs/keys) count"),
lager:info("Run Search URL: ~s", [S#trial_state.solr_url_after]),
verify_count(S#trial_state.solr_url_after, CheckCount).
leave_or_join(Cluster, S=#trial_state{join_node=undefined}) ->
Node = S#trial_state.leave_node,
rt:leave(Node),
?assertEqual(ok, rt:wait_until_unpingable(Node)),
Cluster -- [Node];
leave_or_join(Cluster, S=#trial_state{leave_node=undefined}) ->
Node = S#trial_state.join_node,
NodeAdmin = S#trial_state.admin_node,
ok = rt:start_and_wait(Node),
ok = rt:join(Node, NodeAdmin),
?assertEqual(ok, rt:wait_until_nodes_ready(Cluster)),
?assertEqual(ok, rt:wait_until_no_pending_changes(Cluster)),
Cluster ++ [Node].