riak_test/tests/replication2_pg.erl

1010 lines
36 KiB
Erlang
Raw Normal View History

2013-05-02 16:03:11 +00:00
-module(replication2_pg).
-export([confirm/0]).
-compile(export_all).
-include_lib("eunit/include/eunit.hrl").
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%% Test proxy_get in Default and Advanced mode of 1.3+ repl
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
setup_repl_clusters(Conf) ->
2013-06-07 19:58:03 +00:00
setup_repl_clusters(Conf, false).
setup_repl_clusters(Conf, SSL) ->
2013-05-02 16:03:11 +00:00
NumNodes = 6,
lager:info("Deploy ~p nodes", [NumNodes]),
CertDir = rt_config:get(rt_scratch_dir) ++ "/certs",
2013-06-07 19:58:03 +00:00
%% make a bunch of crypto keys
make_certs:rootCA(CertDir, "rootCA"),
make_certs:intermediateCA(CertDir, "intCA", "rootCA"),
make_certs:endusers(CertDir, "rootCA", ["site3.basho.com", "site4.basho.com"]),
make_certs:endusers(CertDir, "intCA", ["site1.basho.com", "site2.basho.com"]),
2013-06-07 19:58:03 +00:00
SSLConfig1 = [
{riak_core,
[
{ssl_enabled, true},
{certfile, filename:join([CertDir,
"site1.basho.com/cert.pem"])},
{keyfile, filename:join([CertDir,
"site1.basho.com/key.pem"])},
{cacertdir, filename:join([CertDir,
"site1.basho.com/cacerts.pem"])}
2013-06-07 19:58:03 +00:00
]}
],
SSLConfig2 = [
{riak_core,
[
{ssl_enabled, true},
{certfile, filename:join([CertDir,
"site2.basho.com/cert.pem"])},
{keyfile, filename:join([CertDir,
"site2.basho.com/key.pem"])},
{cacertdir, filename:join([CertDir,
"site2.basho.com/cacerts.pem"])}
2013-06-07 19:58:03 +00:00
]}
],
SSLConfig3 = [
{riak_core,
[
{ssl_enabled, true},
{certfile, filename:join([CertDir,
"site3.basho.com/cert.pem"])},
{keyfile, filename:join([CertDir,
"site3.basho.com/key.pem"])},
{cacertdir, filename:join([CertDir,
"site3.basho.com/cacerts.pem"])}
2013-06-07 19:58:03 +00:00
]}
],
2014-04-03 19:15:58 +00:00
rt:set_advanced_conf(all, Conf),
Nodes = [ANodes, BNodes, CNodes] = rt:build_clusters([2, 2, 2]),
2013-06-07 19:58:03 +00:00
rt:wait_for_cluster_service(ANodes, riak_repl),
rt:wait_for_cluster_service(BNodes, riak_repl),
rt:wait_for_cluster_service(CNodes, riak_repl),
2014-04-03 19:15:58 +00:00
AFirst = hd(ANodes),
BFirst = hd(BNodes),
CFirst = hd(CNodes),
2013-05-02 16:03:11 +00:00
rt:log_to_nodes(Nodes, "Starting replication2_pg test"),
lager:info("ANodes: ~p", [ANodes]),
lager:info("BNodes: ~p", [BNodes]),
lager:info("CNodes: ~p", [CNodes]),
2013-06-07 19:58:03 +00:00
case SSL of
true ->
lager:info("Enabling SSL for this test"),
[rt:update_app_config(N, merge_config(SSLConfig1, Conf)) ||
N <- ANodes],
[rt:update_app_config(N, merge_config(SSLConfig2, Conf)) ||
N <- BNodes],
[rt:update_app_config(N, merge_config(SSLConfig3, Conf)) ||
N <- CNodes];
_ ->
lager:info("SSL not enabled for this test")
2013-06-07 19:58:03 +00:00
end,
2013-05-02 16:03:11 +00:00
rt:log_to_nodes(Nodes, "Building and connecting clusters"),
repl_util:name_cluster(AFirst, "A"),
repl_util:name_cluster(BFirst, "B"),
repl_util:name_cluster(CFirst, "C"),
rt:wait_until_ring_converged(ANodes),
rt:wait_until_ring_converged(BNodes),
rt:wait_until_ring_converged(CNodes),
%% get the leader for the first cluster
repl_util:wait_until_leader(AFirst),
LeaderA = rpc:call(AFirst, riak_core_cluster_mgr, get_leader, []),
2014-04-03 19:15:58 +00:00
{ok, {BIP, BPort}} = rpc:call(BFirst, application, get_env,
2013-05-02 16:03:11 +00:00
[riak_core, cluster_mgr]),
2014-04-03 19:15:58 +00:00
repl_util:connect_cluster(LeaderA, BIP, BPort),
2013-05-02 16:03:11 +00:00
2014-04-03 19:15:58 +00:00
{ok, {CIP, CPort}} = rpc:call(CFirst, application, get_env,
2013-05-02 16:03:11 +00:00
[riak_core, cluster_mgr]),
2014-04-03 19:15:58 +00:00
repl_util:connect_cluster(LeaderA, CIP, CPort),
2013-05-02 16:03:11 +00:00
?assertEqual(ok, repl_util:wait_for_connection(LeaderA, "B")),
rt:wait_until_ring_converged(ANodes),
?assertEqual(ok, repl_util:wait_for_connection(LeaderA, "C")),
rt:wait_until_ring_converged(ANodes),
2014-04-03 19:15:58 +00:00
rt:wait_until_transfers_complete(ANodes),
rt:wait_until_transfers_complete(BNodes),
rt:wait_until_transfers_complete(CNodes),
2013-05-02 16:03:11 +00:00
{LeaderA, ANodes, BNodes, CNodes, Nodes}.
make_test_object(Suffix) ->
Bucket = <<"test_bucket">>,
KeyText = "test_key" ++ Suffix,
ValueText = "testdata_" ++ Suffix,
Key = erlang:list_to_binary(KeyText),
Value = erlang:list_to_binary(ValueText),
{Bucket, Key, Value}.
2013-06-07 19:58:03 +00:00
2013-05-02 16:03:11 +00:00
test_basic_pg(Mode) ->
2013-06-07 19:58:03 +00:00
test_basic_pg(Mode, false).
test_basic_pg(Mode, SSL) ->
banner(io_lib:format("test_basic_pg with ~p mode", [Mode]), SSL),
2013-05-02 16:03:11 +00:00
Conf = [
{riak_repl,
[
{proxy_get, enabled},
{fullsync_on_connect, false}
]}
],
2014-04-03 19:15:58 +00:00
{LeaderA, ANodes, BNodes, CNodes, AllNodes} =
2013-06-07 19:58:03 +00:00
setup_repl_clusters(Conf, SSL),
2013-05-02 16:03:11 +00:00
rt:log_to_nodes(AllNodes, "Testing basic pg"),
case Mode of
mode_repl13 ->
ModeRes = rpc:call(LeaderA, riak_repl_console, modes, [["mode_repl13"]]),
lager:info("ModeRes = ~p", [ModeRes]);
mixed ->
lager:info("Using mode_repl12, mode_repl13"),
ok
end,
rt:wait_until_ring_converged(ANodes),
PGEnableResult = rpc:call(LeaderA, riak_repl_console, proxy_get, [["enable","B"]]),
lager:info("Enabled pg: ~p", [PGEnableResult]),
Status = rpc:call(LeaderA, riak_repl_console, status, [quiet]),
case proplists:get_value(proxy_get_enabled, Status) of
undefined -> ?assert(false);
2013-05-02 16:03:11 +00:00
EnabledFor -> lager:info("PG enabled for cluster ~p",[EnabledFor])
end,
PidA = rt:pbc(LeaderA),
{ok,CidA}=riak_repl_pb_api:get_clusterid(PidA),
lager:info("Cluster ID for A = ~p", [CidA]),
{Bucket, KeyA, ValueA} = make_test_object("a"),
{Bucket, KeyB, ValueB} = make_test_object("b"),
rt:pbc_write(PidA, Bucket, KeyA, ValueA),
rt:pbc_write(PidA, Bucket, KeyB, ValueB),
2014-04-03 19:15:58 +00:00
_FirstA = hd(ANodes),
FirstB = hd(BNodes),
FirstC = hd(CNodes),
2013-05-02 16:03:11 +00:00
PidB = rt:pbc(FirstB),
lager:info("Connected to cluster B"),
{ok, PGResult} = riak_repl_pb_api:get(PidB,Bucket,KeyA,CidA),
?assertEqual(ValueA, riakc_obj:get_value(PGResult)),
rt:log_to_nodes(AllNodes, "Disabling pg on A"),
PGDisableResult = rpc:call(LeaderA, riak_repl_console, proxy_get, [["disable","B"]]),
lager:info("Disable pg ~p", [PGDisableResult]),
Status2 = rpc:call(LeaderA, riak_repl_console, status, [quiet]),
?assertEqual([], proplists:get_value(proxy_get_enabled, Status2)),
2013-05-02 16:03:11 +00:00
rt:wait_until_ring_converged(ANodes),
rt:wait_until_ring_converged(BNodes),
%% After the clusters are disconnected, see if the object was
%% written locally after the PG
{ok, PG2Value} = riak_repl_pb_api:get(PidB,Bucket,KeyA,CidA),
?assertEqual(ValueA, riakc_obj:get_value(PG2Value)),
%% test an object that wasn't previously "proxy-gotten", it should fail
FailedResult = riak_repl_pb_api:get(PidB,Bucket,KeyB,CidA),
?assertEqual({error, notfound}, FailedResult),
2013-05-08 16:13:33 +00:00
PGEnableResult2 = rpc:call(LeaderA, riak_repl_console, proxy_get, [["enable","B"]]),
lager:info("Enabled pg: ~p", [PGEnableResult2]),
Status3 = rpc:call(LeaderA, riak_repl_console, status, [quiet]),
case proplists:get_value(proxy_get_enabled, Status3) of
undefined -> ?assert(false);
2013-05-08 16:13:33 +00:00
EnabledFor2 -> lager:info("PG enabled for cluster ~p",[EnabledFor2])
end,
rt:wait_until_ring_converged(ANodes),
rt:wait_until_ring_converged(BNodes),
{ok, PGResult2} = riak_repl_pb_api:get(PidB,Bucket,KeyA,CidA),
?assertEqual(ValueA, riakc_obj:get_value(PGResult2)),
%% Test with optional n_val and sloppy_quorum Options.
%% KeyB is not on C yet. Try via proxy get with above options.
PGEnableResult3 = rpc:call(LeaderA, riak_repl_console, proxy_get, [["enable","C"]]),
lager:info("Enabled pg: ~p", [PGEnableResult3]),
Status4 = rpc:call(LeaderA, riak_repl_console, status, [quiet]),
case proplists:get_value(proxy_get_enabled, Status4) of
undefined -> ?assert(false);
EnabledFor3 -> lager:info("PG enabled for cluster ~p",[EnabledFor3])
end,
PidC = rt:pbc(FirstC),
2013-07-02 13:11:07 +00:00
Options = [{n_val, 1}, {sloppy_quorum, false}],
lager:info("Test proxy get from C using options: ~p", [Options]),
2013-07-02 13:11:07 +00:00
PGResult3 = riak_repl_pb_api:get(PidC,Bucket,KeyA,CidA,Options),
% it's ok if the first request fails due to the options,
% try it again without options to see if it passes
2013-07-02 16:52:32 +00:00
RetriableGet = case PGResult3 of
2013-07-02 13:11:07 +00:00
{ok, PGResult3Value} ->
2013-07-02 16:52:32 +00:00
riakc_obj:get_value(PGResult3Value);
2013-07-02 13:11:07 +00:00
{error, notfound} ->
2013-07-02 16:40:39 +00:00
RetryOptions = [{n_val, 1}],
2013-07-02 17:09:48 +00:00
case riak_repl_pb_api:get(PidC,Bucket,KeyA,CidA,RetryOptions) of
{ok, PGResult4Value} -> riakc_obj:get_value(PGResult4Value);
UnknownResult -> UnknownResult
end;
2013-07-02 13:11:07 +00:00
UnknownResult ->
2013-07-02 16:52:32 +00:00
%% welp, we might have been expecting a notfound, but we got
%% something else.
UnknownResult
2013-07-02 13:11:07 +00:00
end,
2013-07-02 16:52:32 +00:00
?assertEqual(ValueA, RetriableGet),
verify_topology_change(ANodes, BNodes),
2013-05-02 16:03:11 +00:00
pass.
%% test 1.2 replication (aka "Default" repl)
%% Mode is either mode_repl12 or mixed.
2013-05-02 16:03:11 +00:00
%% "mixed" is the default in 1.3: mode_repl12, mode_repl13
test_12_pg(Mode) ->
2013-06-07 19:58:03 +00:00
test_12_pg(Mode, false).
test_12_pg(Mode, SSL) ->
banner(io_lib:format("test_12_pg with ~p mode", [Mode]), SSL),
2013-05-02 16:03:11 +00:00
Conf = [
{riak_repl,
[
{proxy_get, enabled},
{fullsync_on_connect, false}
]}
],
{LeaderA, ANodes, BNodes, CNodes, AllNodes} =
2013-06-07 19:58:03 +00:00
setup_repl_clusters(Conf, SSL),
2013-05-02 16:03:11 +00:00
{Bucket, KeyA, ValueA} = make_test_object("a"),
{Bucket, KeyB, ValueB} = make_test_object("b"),
rt:log_to_nodes(AllNodes, "Test 1.2 proxy_get"),
2014-04-03 19:15:58 +00:00
_FirstA = hd(ANodes),
FirstB = hd(BNodes),
_FirstC = hd(CNodes),
2013-05-02 16:03:11 +00:00
case Mode of
mode_repl12 ->
ModeRes = rpc:call(FirstB, riak_repl_console, modes, [["mode_repl12"]]),
lager:info("ModeRes = ~p", [ModeRes]);
mixed ->
lager:info("Using mode_repl12, mode_repl13"),
ok
end,
[rt:wait_until_ring_converged(Ns) || Ns <- [ANodes, BNodes, CNodes]],
PidA = rt:pbc(LeaderA),
rt:pbc_write(PidA, Bucket, KeyA, ValueA),
rt:pbc_write(PidA, Bucket, KeyB, ValueB),
{ok,CidA}=riak_repl_pb_api:get_clusterid(PidA),
lager:info("Cluster ID for A = ~p", [CidA]),
LeaderB = rpc:call(FirstB, riak_repl2_leader, leader_node, []),
rt:log_to_nodes([LeaderB], "Trying to use PG while it's disabled"),
PidB = rt:pbc(LeaderB),
2013-05-08 16:13:33 +00:00
?assertEqual({error, notfound},
riak_repl_pb_api:get(PidB, Bucket, KeyA, CidA)),
2013-05-02 16:03:11 +00:00
rt:log_to_nodes([LeaderA], "Adding a listener"),
LeaderAIP = rt:get_ip(LeaderA),
ListenerArgs = [[atom_to_list(LeaderA), LeaderAIP, "5666"]],
2013-05-02 16:03:11 +00:00
Res = rpc:call(LeaderA, riak_repl_console, add_listener, ListenerArgs),
?assertEqual(ok, Res),
[rt:wait_until_ring_converged(Ns) || Ns <- [ANodes, BNodes, CNodes]],
rt:log_to_nodes([FirstB], "Adding a site"),
SiteArgs = [LeaderAIP, "5666", "rtmixed"],
2013-05-02 16:03:11 +00:00
Res = rpc:call(FirstB, riak_repl_console, add_site, [SiteArgs]),
lager:info("Res = ~p", [Res]),
rt:log_to_nodes(AllNodes, "Waiting until connected"),
wait_until_12_connection(LeaderA),
[rt:wait_until_ring_converged(Ns) || Ns <- [ANodes, BNodes, CNodes]],
lager:info("Trying proxy_get"),
LeaderB2 = rpc:call(FirstB, riak_repl2_leader, leader_node, []),
PidB2 = rt:pbc(LeaderB2),
{ok, PGResult} = riak_repl_pb_api:get(PidB2, Bucket, KeyB, CidA),
lager:info("PGResult: ~p", [PGResult]),
?assertEqual(ValueB, riakc_obj:get_value(PGResult)),
lager:info("Disable repl and wait for clusters to disconnect"),
rt:log_to_nodes([LeaderA], "Delete listener"),
DelListenerArgs = [[atom_to_list(LeaderA), LeaderAIP, "5666"]],
DelListenerRes = rpc:call(LeaderA, riak_repl_console, del_listener, DelListenerArgs),
?assertEqual(ok, DelListenerRes),
[rt:wait_until_ring_converged(Ns) || Ns <- [ANodes, BNodes, CNodes]],
rt:log_to_nodes([FirstB], "Delete site"),
DelSiteArgs = [LeaderAIP, "5666", "rtmixed"],
DelSiteRes = rpc:call(FirstB, riak_repl_console, add_site, [DelSiteArgs]),
lager:info("Res = ~p", [DelSiteRes]),
rt:log_to_nodes(AllNodes, "Waiting until disconnected"),
wait_until_12_no_connection(LeaderA),
[rt:wait_until_ring_converged(Ns) || Ns <- [ANodes, BNodes, CNodes]],
rt:log_to_nodes(AllNodes, "Trying proxy_get without a connection"),
?assertEqual({error, notfound},
riak_repl_pb_api:get(PidB, Bucket, KeyA, CidA)),
2013-05-02 16:03:11 +00:00
pass.
%% test shutting down nodes in source + sink clusters
test_pg_proxy() ->
2013-06-07 19:58:03 +00:00
test_pg_proxy(false).
test_pg_proxy(SSL) ->
banner("test_pg_proxy", SSL),
2013-05-02 16:03:11 +00:00
Conf = [
{riak_repl,
[
{proxy_get, enabled},
{fullsync_on_connect, false}
]}
],
2014-04-03 19:15:58 +00:00
{LeaderA, ANodes, BNodes, CNodes, AllNodes} =
2013-06-07 19:58:03 +00:00
setup_repl_clusters(Conf, SSL),
2013-05-02 16:03:11 +00:00
rt:log_to_nodes(AllNodes, "Testing pg proxy"),
rt:wait_until_ring_converged(ANodes),
PGEnableResult = rpc:call(LeaderA, riak_repl_console, proxy_get, [["enable","B"]]),
lager:info("Enabled pg: ~p", [PGEnableResult]),
Status = rpc:call(LeaderA, riak_repl_console, status, [quiet]),
case proplists:get_value(proxy_get_enabled, Status) of
undefined -> ?assert(false);
2013-05-02 16:03:11 +00:00
EnabledFor -> lager:info("PG enabled for cluster ~p",[EnabledFor])
end,
PidA = rt:pbc(LeaderA),
{ok,CidA}=riak_repl_pb_api:get_clusterid(PidA),
lager:info("Cluster ID for A = ~p", [CidA]),
%% Write a new k/v for every PG test, otherwise you'll get a locally written response
{Bucket, KeyA, ValueA} = make_test_object("a"),
{Bucket, KeyB, ValueB} = make_test_object("b"),
{Bucket, KeyC, ValueC} = make_test_object("c"),
{Bucket, KeyD, ValueD} = make_test_object("d"),
rt:pbc_write(PidA, Bucket, KeyA, ValueA),
rt:pbc_write(PidA, Bucket, KeyB, ValueB),
rt:pbc_write(PidA, Bucket, KeyC, ValueC),
rt:pbc_write(PidA, Bucket, KeyD, ValueD),
%% sanity check. You know, like the 10000 tests that autoconf runs
2013-05-02 16:03:11 +00:00
%% before it actually does any work.
2014-04-03 19:15:58 +00:00
FirstA = hd(ANodes),
FirstB = hd(BNodes),
_FirstC = hd(CNodes),
2013-05-02 16:03:11 +00:00
PidB = rt:pbc(FirstB),
lager:info("Connected to cluster B"),
{ok, PGResult} = riak_repl_pb_api:get(PidB,Bucket,KeyA,CidA),
?assertEqual(ValueA, riakc_obj:get_value(PGResult)),
rt:wait_until_transfers_complete(ANodes),
rt:wait_until_transfers_complete(BNodes),
2013-05-02 16:03:11 +00:00
lager:info("Stopping leader on requester cluster"),
PGLeaderB = rpc:call(FirstB, riak_core_cluster_mgr, get_leader, []),
rt:log_to_nodes(AllNodes, "Killing leader on requester cluster"),
rt:stop(PGLeaderB),
[RunningBNode | _ ] = BNodes -- [PGLeaderB],
repl_util:wait_until_leader(RunningBNode),
PidB2 = rt:pbc(RunningBNode),
2013-05-02 16:03:11 +00:00
lager:info("Now trying proxy_get"),
?assertEqual(ok, wait_until_pg(RunningBNode, PidB2, Bucket, KeyC, CidA)),
2013-05-02 16:03:11 +00:00
lager:info("If you got here, proxy_get worked after the pg block requesting leader was killed"),
lager:info("Stopping leader on provider cluster"),
PGLeaderA = rpc:call(FirstA, riak_core_cluster_mgr, get_leader, []),
rt:stop(PGLeaderA),
[RunningANode | _ ] = ANodes -- [PGLeaderA],
repl_util:wait_until_leader(RunningANode),
?assertEqual(ok, wait_until_pg(RunningBNode, PidB2, Bucket, KeyD, CidA)),
2013-05-02 16:03:11 +00:00
lager:info("If you got here, proxy_get worked after the pg block providing leader was killed"),
lager:info("pg_proxy test complete. Time to obtain celebratory cheese sticks."),
pass.
%% test mapping of cluster from a retired cluster to an active one, repl issue 306
test_cluster_mapping() ->
test_cluster_mapping(false).
test_cluster_mapping(SSL) ->
banner("test_cluster_mapping", SSL),
Conf = [
{riak_repl,
[
{proxy_get, enabled},
{fullsync_on_connect, false}
]}
],
2014-04-03 19:15:58 +00:00
{LeaderA, ANodes, BNodes, CNodes, _AllNodes} =
setup_repl_clusters(Conf, SSL),
2014-04-03 19:15:58 +00:00
_FirstA = hd(ANodes),
FirstB = hd(BNodes),
FirstC = hd(CNodes),
LeaderB = rpc:call(FirstB, riak_core_cluster_mgr, get_leader, []),
LeaderC = rpc:call(FirstC, riak_core_cluster_mgr, get_leader, []),
% Cluser C-> connection must be set up for the proxy gets to work
% with the cluster ID mapping
{ok, {CIP, CPort}} = rpc:call(FirstC, application, get_env,
[riak_core, cluster_mgr]),
repl_util:connect_cluster(LeaderB, CIP, CPort),
?assertEqual(ok, repl_util:wait_for_connection(LeaderB, "C")),
% enable A to serve blocks to C
PGEnableResultA = rpc:call(LeaderA, riak_repl_console, proxy_get, [["enable","C"]]),
% enable B to serve blocks to C
PGEnableResultB = rpc:call(LeaderB, riak_repl_console, proxy_get, [["enable","C"]]),
lager:info("Enabled pg to A:~p", [PGEnableResultA]),
lager:info("Enabled pg to B:~p", [PGEnableResultB]),
StatusA = rpc:call(LeaderA, riak_repl_console, status, [quiet]),
case proplists:get_value(proxy_get_enabled, StatusA) of
undefined -> ?assert(false);
EnabledForA -> lager:info("PG enabled for cluster ~p",[EnabledForA])
end,
StatusB = rpc:call(LeaderB, riak_repl_console, status, [quiet]),
case proplists:get_value(proxy_get_enabled, StatusB) of
undefined -> ?assert(false);
EnabledForB -> lager:info("PG enabled for cluster ~p",[EnabledForB])
end,
[rt:wait_until_ring_converged(Ns) || Ns <- [ANodes, BNodes, CNodes]],
PidA = rt:pbc(LeaderA),
{ok,CidA}=riak_repl_pb_api:get_clusterid(PidA),
lager:info("Cluster ID for A = ~p", [CidA]),
PidB = rt:pbc(LeaderB),
{ok,CidB}=riak_repl_pb_api:get_clusterid(PidB),
lager:info("Cluster ID for B = ~p", [CidB]),
PidC = rt:pbc(LeaderC),
{ok,CidC}=riak_repl_pb_api:get_clusterid(PidC),
lager:info("Cluster ID for C = ~p", [CidC]),
%% Write a new k/v for every PG test, otherwise you'll get a locally written response
{Bucket, KeyA, ValueA} = make_test_object("a"),
{Bucket, KeyB, ValueB} = make_test_object("b"),
{Bucket, KeyC, ValueC} = make_test_object("c"),
{Bucket, KeyD, ValueD} = make_test_object("d"),
rt:pbc_write(PidA, Bucket, KeyA, ValueA),
rt:pbc_write(PidA, Bucket, KeyB, ValueB),
rt:pbc_write(PidA, Bucket, KeyC, ValueC),
rt:pbc_write(PidA, Bucket, KeyD, ValueD),
{ok, PGResult} = riak_repl_pb_api:get(PidA,Bucket,KeyA,CidA),
?assertEqual(ValueA, riakc_obj:get_value(PGResult)),
% Configure cluster_mapping on C to map cluster_id A -> C
lager:info("Configuring cluster C to map its cluster_id to B's cluster_id"),
%rpc:call(LeaderC, riak_core_metadata, put, [{<<"replication">>, <<"cluster-mapping">>}, CidA, CidB]),
2013-09-12 01:54:22 +00:00
rpc:call(LeaderC, riak_repl_console, add_block_provider_redirect, [[CidA, CidB]]),
Res = rpc:call(LeaderC, riak_core_metadata, get, [{<<"replication">>, <<"cluster-mapping">>}, CidA]),
lager:info("result: ~p", [Res]),
% full sync from CS Block Provider A to CS Block Provider B
repl_util:enable_fullsync(LeaderA, "B"),
rt:wait_until_ring_converged(ANodes),
{Time,_} = timer:tc(repl_util,start_and_wait_until_fullsync_complete,[LeaderA]),
lager:info("Fullsync completed in ~p seconds", [Time/1000/1000]),
% shut down cluster A
lager:info("Shutting down cluster A"),
[ rt:stop(Node) || Node <- ANodes ],
[ rt:wait_until_unpingable(Node) || Node <- ANodes ],
rt:wait_until_ring_converged(BNodes),
rt:wait_until_ring_converged(CNodes),
% proxy-get from cluster C, using A's clusterID
% Should redirect requester C from Cid A, to Cid B, and still
% return the correct value for the Key
{ok, PGResultC} = riak_repl_pb_api:get(PidC, Bucket, KeyC, CidA),
lager:info("PGResultC: ~p", [PGResultC]),
?assertEqual(ValueC, riakc_obj:get_value(PGResultC)),
% now delete the redirect and make sure it's gone
rpc:call(LeaderC, riak_repl_console, delete_block_provider_redirect, [[CidA]]),
case rpc:call(LeaderC, riak_core_metadata, get, [{<<"replication">>, <<"cluster-mapping">>}, CidA]) of
undefined ->
2013-09-12 01:54:22 +00:00
lager:info("cluster-mapping no longer found in meta data, after delete, which is expected");
Match ->
lager:info("cluster mapping ~p still in meta data after delete; problem!", [Match]),
?assert(false)
end,
pass.
2013-05-02 16:03:11 +00:00
%% connect source + sink clusters, pg bidirectionally
test_bidirectional_pg() ->
2013-06-11 15:33:03 +00:00
test_bidirectional_pg(false).
2013-06-07 19:58:03 +00:00
test_bidirectional_pg(SSL) ->
banner("test_bidirectional_pg", SSL),
2013-05-02 16:03:11 +00:00
Conf = [
{riak_repl,
[
{proxy_get, enabled},
{fullsync_on_connect, false}
]}
],
2014-04-03 19:15:58 +00:00
{LeaderA, ANodes, BNodes, CNodes, AllNodes} =
2013-06-07 19:58:03 +00:00
setup_repl_clusters(Conf, SSL),
2013-05-02 16:03:11 +00:00
rt:log_to_nodes(AllNodes, "Testing bidirectional proxy-get"),
rt:wait_until_ring_converged(ANodes),
rt:wait_until_ring_converged(BNodes),
2014-04-03 19:15:58 +00:00
FirstA = hd(ANodes),
FirstB = hd(BNodes),
_FirstC = hd(CNodes),
2013-05-02 16:03:11 +00:00
LeaderB = rpc:call(FirstB, riak_repl2_leader, leader_node, []),
{ok, {AIP, APort}} = rpc:call(FirstA, application, get_env,
2013-05-02 16:03:11 +00:00
[riak_core, cluster_mgr]),
repl_util:connect_cluster(LeaderB, AIP, APort),
2013-05-02 16:03:11 +00:00
rt:wait_until_ring_converged(ANodes),
rt:wait_until_ring_converged(BNodes),
PGEnableResult = rpc:call(LeaderA, riak_repl_console, proxy_get, [["enable","B"]]),
PGEnableResult = rpc:call(LeaderB, riak_repl_console, proxy_get, [["enable","A"]]),
2013-05-02 16:03:11 +00:00
lager:info("Enabled bidirectional pg ~p", [PGEnableResult]),
StatusA = rpc:call(LeaderA, riak_repl_console, status, [quiet]),
case proplists:get_value(proxy_get_enabled, StatusA) of
undefined -> ?assert(false);
2013-05-02 16:03:11 +00:00
EnabledForA -> lager:info("PG enabled for cluster ~p",[EnabledForA])
end,
StatusB = rpc:call(LeaderB, riak_repl_console, status, [quiet]),
case proplists:get_value(proxy_get_enabled, StatusB) of
undefined -> ?assert(false);
2013-05-02 16:03:11 +00:00
EnabledForB -> lager:info("PG enabled for cluster ~p",[EnabledForB])
end,
PidA = rt:pbc(LeaderA),
PidB = rt:pbc(FirstB),
{ok,CidA}=riak_repl_pb_api:get_clusterid(PidA),
{ok,CidB}=riak_repl_pb_api:get_clusterid(PidB),
lager:info("Cluster ID for A = ~p", [CidA]),
lager:info("Cluster ID for B = ~p", [CidB]),
{Bucket, KeyA, ValueA} = make_test_object("a"),
{Bucket, KeyB, ValueB} = make_test_object("b"),
%% write some data to cluster A
rt:pbc_write(PidA, Bucket, KeyA, ValueA),
%% write some data to cluster B
rt:pbc_write(PidB, Bucket, KeyB, ValueB),
lager:info("Trying first get"),
wait_until_pg(LeaderB, PidB, Bucket, KeyA, CidA),
lager:info("First get worked"),
lager:info("Trying second get"),
wait_until_pg(LeaderA, PidA, Bucket, KeyB, CidB),
lager:info("Second get worked"),
verify_topology_change(ANodes, BNodes),
2013-05-02 16:03:11 +00:00
pass.
%% Test multiple sinks against a single source
test_multiple_sink_pg() ->
2013-06-07 19:58:03 +00:00
test_multiple_sink_pg(false).
test_multiple_sink_pg(SSL) ->
banner("test_multiple_sink_pg", SSL),
2013-05-02 16:03:11 +00:00
Conf = [
{riak_repl,
[
{proxy_get, enabled},
{fullsync_on_connect, false}
]}
],
{LeaderA, ANodes, BNodes, CNodes, AllNodes} =
2013-06-07 19:58:03 +00:00
setup_repl_clusters(Conf, SSL),
2013-05-02 16:03:11 +00:00
rt:log_to_nodes(AllNodes, "Testing basic pg"),
rt:wait_until_ring_converged(ANodes),
rt:wait_until_ring_converged(BNodes),
rt:wait_until_ring_converged(CNodes),
PGEnableResultB = rpc:call(LeaderA, riak_repl_console, proxy_get, [["enable","B"]]),
PGEnableResultC = rpc:call(LeaderA, riak_repl_console, proxy_get, [["enable","C"]]),
lager:info("Enabled pg to B:~p", [PGEnableResultB]),
lager:info("Enabled pg to C:~p", [PGEnableResultC]),
Status = rpc:call(LeaderA, riak_repl_console, status, [quiet]),
case proplists:get_value(proxy_get_enabled, Status) of
undefined -> ?assert(false);
2013-05-02 16:03:11 +00:00
EnabledForC -> lager:info("PG enabled for cluster ~p",[EnabledForC])
end,
PidA = rt:pbc(LeaderA),
{ok,CidA}=riak_repl_pb_api:get_clusterid(PidA),
lager:info("Cluster ID for A = ~p", [CidA]),
{Bucket, KeyA, ValueA} = make_test_object("a"),
{Bucket, KeyB, ValueB} = make_test_object("b"),
rt:pbc_write(PidA, Bucket, KeyA, ValueA),
rt:pbc_write(PidA, Bucket, KeyB, ValueB),
2014-04-03 19:15:58 +00:00
_FirstA = hd(ANodes),
FirstB = hd(BNodes),
FirstC = hd(CNodes),
2013-05-02 16:03:11 +00:00
PidB = rt:pbc(FirstB),
PidC = rt:pbc(FirstC),
{ok, PGResultB} = riak_repl_pb_api:get(PidB,Bucket,KeyA,CidA),
?assertEqual(ValueA, riakc_obj:get_value(PGResultB)),
{ok, PGResultC} = riak_repl_pb_api:get(PidC,Bucket,KeyB,CidA),
?assertEqual(ValueB, riakc_obj:get_value(PGResultC)),
pass.
%% test 1.2 + 1.3 repl being used at the same time
test_mixed_pg() ->
2013-06-07 19:58:03 +00:00
test_mixed_pg(false).
test_mixed_pg(SSL) ->
banner("test_mixed_pg", SSL),
2013-05-02 16:03:11 +00:00
Conf = [
{riak_repl,
[
{proxy_get, enabled},
{fullsync_on_connect, false}
]}
],
{LeaderA, ANodes, BNodes, CNodes, AllNodes} =
2013-06-07 19:58:03 +00:00
setup_repl_clusters(Conf, SSL),
2013-05-02 16:03:11 +00:00
rt:log_to_nodes(AllNodes, "Testing basic pg"),
rt:wait_until_ring_converged(ANodes),
PGEnableResult = rpc:call(LeaderA, riak_repl_console, proxy_get, [["enable","B"]]),
lager:info("Enabled pg: ~p", [PGEnableResult]),
2013-05-02 16:03:11 +00:00
Status = rpc:call(LeaderA, riak_repl_console, status, [quiet]),
case proplists:get_value(proxy_get_enabled, Status) of
undefined -> ?assert(false);
2013-05-02 16:03:11 +00:00
EnabledFor -> lager:info("PG enabled for cluster ~p",[EnabledFor])
end,
PidA = rt:pbc(LeaderA),
{ok,CidA}=riak_repl_pb_api:get_clusterid(PidA),
lager:info("Cluster ID for A = ~p", [CidA]),
{Bucket, KeyB, ValueB} = make_test_object("b"),
{Bucket, KeyC, ValueC} = make_test_object("c"),
rt:pbc_write(PidA, Bucket, KeyB, ValueB),
rt:pbc_write(PidA, Bucket, KeyC, ValueC),
2014-04-03 19:15:58 +00:00
_FirstA = hd(ANodes),
FirstB = hd(BNodes),
FirstC = hd(CNodes),
2013-05-02 16:03:11 +00:00
rt:wait_until_ring_converged(ANodes),
rt:wait_until_ring_converged(BNodes),
rt:log_to_nodes([LeaderA], "Adding a listener"),
ListenerIP = rt:get_ip(LeaderA),
ListenerArgs = [[atom_to_list(LeaderA), ListenerIP, "5666"]],
2013-05-02 16:03:11 +00:00
Res = rpc:call(LeaderA, riak_repl_console, add_listener, ListenerArgs),
?assertEqual(ok, Res),
[rt:wait_until_ring_converged(Ns) || Ns <- [ANodes, BNodes, CNodes]],
rt:log_to_nodes([FirstC], "Adding a site"),
SiteArgs = [ListenerIP, "5666", "rtmixed"],
2013-05-02 16:03:11 +00:00
Res = rpc:call(FirstC, riak_repl_console, add_site, [SiteArgs]),
lager:info("Res = ~p", [Res]),
rt:log_to_nodes(AllNodes, "Waiting until connected"),
wait_until_12_connection(LeaderA),
[rt:wait_until_ring_converged(Ns) || Ns <- [ANodes, BNodes, CNodes]],
lager:info("Trying proxy_get"),
LeaderC = rpc:call(FirstC, riak_repl2_leader, leader_node, []),
PidB = rt:pbc(FirstB),
PidC = rt:pbc(LeaderC),
{ok, PGResultB} = riak_repl_pb_api:get(PidB, Bucket, KeyB, CidA),
lager:info("PGResultB: ~p", [PGResultB]),
?assertEqual(ValueB, riakc_obj:get_value(PGResultB)),
{ok, PGResultC} = riak_repl_pb_api:get(PidC, Bucket, KeyC, CidA),
lager:info("PGResultC: ~p", [PGResultC]),
?assertEqual(ValueC, riakc_obj:get_value(PGResultC)),
pass.
wait_until_12_connection(Node) ->
rt:wait_until(Node,
fun(_) ->
case rpc:call(Node, riak_repl_console, status, [quiet]) of
{badrpc, _} ->
false;
Status ->
case proplists:get_value(server_stats, Status) of
[] ->
false;
[{_, _, too_busy}] ->
false;
[_C] ->
true;
Conns ->
lager:warning("multiple connections detected: ~p",
[Conns]),
true
end
end
end). %% 40 seconds is enough for repl
wait_until_12_no_connection(Node) ->
rt:wait_until(Node,
fun(_) ->
case rpc:call(Node, riak_repl_console, status, [quiet]) of
{badrpc, _} ->
false;
Status ->
case proplists:get_value(server_stats, Status) of
undefined ->
true;
[] ->
true;
[{_, _, too_busy}] ->
false;
[_C] ->
false;
Conns ->
lager:warning("multiple connections detected: ~p",
[Conns]),
false
end
end
end). %% 40 seconds is enough for repl
2013-05-02 16:03:11 +00:00
%% these funs allow you to call:
%% riak_test -t replication2_pg:test_basic_pg_mode_repl13 etc
test_basic_pg_mode_repl13() ->
test_basic_pg(mode_repl13).
test_basic_pg_mode_mixed() ->
test_basic_pg(mixed).
test_12_pg_mode_repl12() ->
test_12_pg(mode_repl12).
test_12_pg_mode_repl_mixed() ->
test_12_pg(mixed).
2013-06-07 19:58:03 +00:00
test_basic_pg_mode_repl13_ssl() ->
test_basic_pg(mode_repl13, true).
test_basic_pg_mode_mixed_ssl() ->
test_basic_pg(mixed, true).
test_12_pg_mode_repl12_ssl() ->
test_12_pg(mode_repl12, true).
test_12_pg_mode_repl_mixed_ssl() ->
test_12_pg(mixed, true).
test_mixed_pg_ssl() ->
test_mixed_pg(true).
test_multiple_sink_pg_ssl() ->
test_multiple_sink_pg(true).
test_bidirectional_pg_ssl() ->
test_bidirectional_pg(true).
test_pg_proxy_ssl() ->
test_pg_proxy(true).
2013-05-02 16:03:11 +00:00
confirm() ->
AllTests =
[
test_basic_pg_mode_repl13,
test_basic_pg_mode_mixed,
test_12_pg_mode_repl12,
test_12_pg_mode_repl_mixed,
test_mixed_pg,
test_multiple_sink_pg,
test_bidirectional_pg,
test_cluster_mapping,
2013-06-07 19:58:03 +00:00
test_pg_proxy,
test_basic_pg_mode_repl13_ssl,
test_basic_pg_mode_mixed_ssl,
test_12_pg_mode_repl12_ssl,
test_12_pg_mode_repl_mixed_ssl,
test_mixed_pg_ssl,
test_multiple_sink_pg_ssl,
test_bidirectional_pg_ssl,
test_pg_proxy_ssl
2013-05-02 16:03:11 +00:00
],
lager:error("run riak_test with -t Mod:test1 -t Mod:test2"),
lager:error("The runnable tests in this module are: ~p", [AllTests]),
2014-04-14 17:55:49 +00:00
%% TODO: The problem with this LC is that it doesn't incorporate any
%% of riak_test's setup/teardown per test.
[?assertEqual(pass, erlang:apply(?MODULE, Test, [])) || Test <- AllTests].
2013-06-07 19:58:03 +00:00
2013-05-02 16:03:11 +00:00
banner(T) ->
2013-06-07 19:58:03 +00:00
banner(T, false).
banner(T, SSL) ->
2013-05-02 16:03:11 +00:00
lager:info("----------------------------------------------"),
lager:info("----------------------------------------------"),
2013-06-07 19:58:03 +00:00
lager:info("~s, SSL ~s",[T, SSL]),
2013-05-02 16:03:11 +00:00
lager:info("----------------------------------------------"),
lager:info("----------------------------------------------").
wait_until_pg(Node, Pid, Bucket, Key, Cid) ->
rt:wait_until(Node,
fun(_) ->
case riak_repl_pb_api:get(Pid,Bucket,Key,Cid) of
{error, notfound} ->
false;
{ok, _Value} -> true;
_ -> false
end
end).
2013-06-07 19:58:03 +00:00
merge_config(Mixin, Base) ->
lists:ukeymerge(1, lists:keysort(1, Mixin), lists:keysort(1, Base)).
verify_topology_change(SourceNodes, SinkNodes) ->
lager:info("Verify topology changes doesn't break the proxy get."),
%% Get connections
[SourceNode1, _SourceNode2] = SourceNodes,
SourceNode1Pid = rt:pbc(SourceNode1),
[SinkNode1, SinkNode2] = SinkNodes,
SinkNode1Pid = rt:pbc(SinkNode1),
{ok, SourceCid} = riak_repl_pb_api:get_clusterid(SourceNode1Pid),
%% Write new object to source.
lager:info("Writing key 'before' to the source."),
{Bucket, KeyBefore, ValueBefore} = make_test_object("before"),
rt:pbc_write(SourceNode1Pid, Bucket, KeyBefore, ValueBefore),
%% Verify proxy_get through the sink works.
lager:info("Verifying key 'before' can be read through the sink."),
{ok, PGResult1} = riak_repl_pb_api:get(SinkNode1Pid,
Bucket, KeyBefore, SourceCid),
?assertEqual(ValueBefore, riakc_obj:get_value(PGResult1)),
%% Remove leader from the sink cluster.
SinkLeader = rpc:call(SinkNode1,
riak_repl2_leader, leader_node, []),
2014-04-14 17:55:49 +00:00
%% Sad this takes 2.5 minutes
lager:info("Removing current leader from the cluster: ~p.",
[SinkLeader]),
rt:leave(SinkLeader),
?assertEqual(ok, rt:wait_until_unpingable(SinkLeader)),
%% Wait for everything to restart, and rings to converge.
lager:info("Starting leader node back up and waiting for repl."),
rt:start(SinkLeader),
rt:wait_for_service(SinkLeader, riak_repl),
rt:wait_until_ring_converged(SinkNodes),
%% Assert nodes have different leaders, which are themselves.
lager:info("Ensure that each node is its own leader."),
SinkNode1Leader = rpc:call(SinkNode1,
riak_repl2_leader, leader_node, []),
SinkNode2Leader = rpc:call(SinkNode2,
riak_repl2_leader, leader_node, []),
?assertEqual(SinkNode1, SinkNode1Leader),
?assertEqual(SinkNode2, SinkNode2Leader),
?assertNotEqual(SinkNode1Leader, SinkNode2Leader),
%% Before we join the nodes, install an intercept on all nodes for
%% the leader election callback.
lager:info("Installing set_leader_node intercept."),
2014-04-14 17:55:49 +00:00
Result = riak_repl2_leader_intercepts:set_leader_node(SinkLeader),
lager:info("riak_repl2_leader_intercepts:set_leader_node(~p) = ~p", [SinkLeader, Result]),
[ begin
rt_intercept:load_code(N),
ok = rt_intercept:add(N, {riak_repl2_leader, [{{set_leader,3}, set_leader_node}]})
end || N <- SinkNodes ],
%% Restart former leader and rejoin to the cluster.
lager:info("Rejoining former leader."),
case SinkLeader of
SinkNode1 ->
rt:join(SinkNode1, SinkNode2);
SinkNode2 ->
rt:join(SinkNode2, SinkNode1)
end,
rt:wait_until_ring_converged(SinkNodes),
%% Assert that all nodes have the same leader.
lager:info("Assert that all nodes have the same leader."),
SinkNode1LeaderRejoin = rpc:call(SinkNode1,
riak_repl2_leader, leader_node, []),
SinkNode2LeaderRejoin = rpc:call(SinkNode2,
riak_repl2_leader, leader_node, []),
?assertEqual(SinkNode1LeaderRejoin, SinkNode2LeaderRejoin),
%% Assert that the leader is the former leader.
lager:info("Assert that new leader is the former leader."),
?assertEqual(SinkLeader, SinkNode1LeaderRejoin),
%% Write new object to source.
lager:info("Writing key 'after' to the source."),
{ok, SourceCid} = riak_repl_pb_api:get_clusterid(SourceNode1Pid),
{Bucket, KeyPost, ValuePost} = make_test_object("after"),
rt:pbc_write(SourceNode1Pid, Bucket, KeyPost, ValuePost),
%% Verify we can retrieve from source.
lager:info("Verifying key 'after' can be read through the source."),
{ok, PGResult2} = riak_repl_pb_api:get(SourceNode1Pid,
Bucket, KeyPost, SourceCid),
?assertEqual(ValuePost, riakc_obj:get_value(PGResult2)),
%% Verify proxy_get through the sink works.
lager:info("Verifying key 'after' can be read through the sink."),
wait_until_pg(SinkNode1, SinkNode1Pid, Bucket, KeyPost, SourceCid),
%% We're good!
pass.