forked from Netflix/dynomite
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathdyno_cluster.py
216 lines (184 loc) · 8.41 KB
/
dyno_cluster.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
#!/usr/bin/env python3
from collections import namedtuple
from plumbum import local
import os
import redis
import random
import shutil
import signal
import yaml
from dyno_node import DynoNode
from utils import *
# TODO: Make this path absolute based on param instead of relative.
CLUSTER_DESC_FILEPATH='../running_cluster.yaml'
DYN_O_MITE_DEFAULTS = dict(
secure_server_option='datacenter',
pem_key_file='conf/dynomite.pem',
data_store=0,
datastore_connections=1,
)
INTERNODE_LISTEN = 8101
CLIENT_LISTEN = 8102
REDIS_PORT = 1212
STATS_PORT = 22222
class DynoSpec(namedtuple('DynoSpec', 'ip dnode_port client_port rack dc token '
'local_connections remote_connections seed_string req_conf')):
"""Specifies how to launch a dynomite node"""
def __new__(cls, ip, dnode_port, client_port, rack, dc, token, local_connections,
remote_connections, req_conf):
seed_string = '{}:{}:{}:{}:{}'.format(ip, dnode_port, rack, dc, token)
return super(DynoSpec, cls).__new__(cls, ip, dnode_port, client_port, rack,
dc, token, local_connections, remote_connections, seed_string, req_conf)
def __init__(self, ip, dnode_port, client_port, rack, dc, token, local_connections,
remote_connections, req_conf):
self.data_store_port = REDIS_PORT
self.stats_port = STATS_PORT
def _generate_config(self, seeds_list):
conf = dict(DYN_O_MITE_DEFAULTS)
conf['datacenter'] = self.dc
conf['rack'] = self.rack
dyn_listen = '{}:{}'.format(self.ip, self.dnode_port)
conf['dyn_listen'] = dyn_listen
conf['listen'] = '{}:{}'.format(self.ip, self.client_port)
# filter out our own seed string
conf['dyn_seeds'] = [s for s in seeds_list if s != self.seed_string]
conf['servers'] = ['{}:{}:0'.format(self.ip, self.data_store_port)]
conf['stats_listen'] = '{}:{}'.format(self.ip, self.stats_port)
conf['tokens'] = self.token
conf['local_peer_connections'] = self.local_connections
conf['remote_peer_connections'] = self.remote_connections
# Add configurations based on the request.
for conf_key, conf_value in self.req_conf.items():
conf[conf_key] = conf_value
return dict(dyn_o_mite=conf)
def write_config(self, seeds_list):
config = self._generate_config(seeds_list)
filename = 'conf/{}:{}:{}.yml'.format(self.dc, self.rack, self.token)
with open(filename, 'w') as fh:
yaml.dump(config, fh, default_flow_style=False)
return filename
class DynoCluster(object):
def __init__(self, request_file, ips):
# Load the YAML file describing the cluster.
with open(request_file, 'r') as fh:
self.request = yaml.load(fh)
self.ips = ips
self.nodes = []
self.counts_by_dc = {}
self.counts_by_rack = {}
# Generate the specification for each node to be started in the cluster.
self.specs = list(self._generate_dynomite_specs())
def _generate_dynomite_specs(self):
tokens = tokens_for_cluster(self.request['cluster_desc'], None)
self.counts_by_rack = dict_request(self.request['cluster_desc'], 'name', 'racks')
self.counts_by_dc = sum_racks(self.counts_by_rack)
total_nodes = sum(self.counts_by_dc.values())
for dc, racks in tokens:
dc_count = self.counts_by_dc[dc]
rack_count = self.counts_by_rack[dc]
remote_count = total_nodes - dc_count
for rack, tokens in racks:
local_count = rack_count[rack] - 1
for token in tokens:
ip = next(self.ips)
yield DynoSpec(ip, INTERNODE_LISTEN, CLIENT_LISTEN, rack, dc, token,
local_count, remote_count, self.request['conf'])
def _get_cluster_desc_yaml(self):
yaml_desc = dict(test_dir=str(local.cwd))
cluster_desc = [dict(name='dyno_nodes')]
cluster_desc.append(dict(name='redis_nodes'))
cluster_desc[0]['pids']=[]
cluster_desc[1]['pids']=[]
for node in self.nodes:
cluster_desc[0]['pids'].append(node.get_dyno_node_pid())
cluster_desc[1]['pids'].append(node.get_storage_node_pid())
yaml_desc['cluster_desc'] = cluster_desc
return yaml_desc
def _pre_launch_sanity_check(self):
"""Checks if there is a cluster already running and tears it down"""
teardown_running_cluster(CLUSTER_DESC_FILEPATH)
def _write_running_cluster_file(self):
yaml_cluster_desc = self._get_cluster_desc_yaml()
with open(CLUSTER_DESC_FILEPATH, 'w') as outfile:
yaml.dump(yaml_cluster_desc, outfile, default_flow_style=False)
def _print_cluster_topology(self):
tokens = tokens_for_cluster(self.request['cluster_desc'], None)
print("Cluster topology:-");
for dc, racks in tokens:
print("\tDC: %s" % dc)
for rack, tokens in racks:
print("\t\tRack: %s" % rack)
# Nested loop is okay here since the #nodes will always be small.
for node in self.nodes:
if node.spec.dc == dc and node.spec.rack == rack:
print("\t\t\tNode: %s || PID: %s" % (node.name, \
node.get_dyno_node_pid()))
def _delete_running_cluster_file(self):
os.remove(CLUSTER_DESC_FILEPATH)
def launch(self):
self._pre_launch_sanity_check()
# Get the list of seeds from the specification for each node.
seeds_list = [spec.seed_string for spec in self.specs]
# Launch each individual Dyno node.
self.nodes = [DynoNode(spec, seeds_list) for spec in self.specs]
for n in self.nodes:
n.launch()
# Now that the cluster is up, write its description to a file.
self._write_running_cluster_file()
self._print_cluster_topology()
def teardown(self):
for n in self.nodes:
n.teardown()
# Delete the cluster description file if it exists.
self._delete_running_cluster_file()
def __enter__(self):
self.launch()
def __exit__(self, type_, value, traceback):
self.teardown()
def enable_read_repairs(self):
for node in self.nodes:
r = make_get_rest_call('http://%s:22222/read_repairs/enable' % node.ip)
assert r.text.find('ENABLED') != -1
def disable_read_repairs(self):
for node in self.nodes:
r = make_get_rest_call('http://%s:22222/read_repairs/disable' % node.ip)
assert r.text.find('DISABLED') != -1
def set_cluster_consistency_level(self, quorum_option):
assert "DC_ONE" in quorum_option or \
"DC_QUORUM" in quorum_option or \
"DC_SAFE_QUORUM" in quorum_option
for node in self.nodes:
r = make_get_rest_call('http://%s:22222/set_consistency/read/%s' % \
(node.ip, quorum_option))
r = make_get_rest_call('http://%s:22222/set_consistency/write/%s' % \
(node.ip, quorum_option))
# Returns the name of the first DC with multiple racks along with the rack count.
# Returns None if no DC has multiple racks.
def get_multi_rack_dc(self):
for dc, racks in self.counts_by_rack.items():
if len(racks) > 1:
return dc, racks
# Returns the DynoNode object of the node that contains 'key' under 'dc' and 'rack'.
def find_node_with_key(self, dc, rack, key):
for node in self.nodes:
if node.spec.dc == dc and node.spec.rack == rack:
node_conn = node.get_data_store_connection()
if node_conn.exists(key):
return node
# Checks if 'node' is part of a DC that has multiple racks.
def _is_node_in_multi_rack_dc(self, node):
source_dc = node.spec.dc
for dc, racks in self.counts_by_rack.items():
if source_dc == dc and len(racks) > 1:
return True
return False
def get_connection(self):
node = random.choice(self.nodes)
return node.get_connection()
def get_connection_to_multi_rack_dc(self):
# Attempt this an arbitrary number of times, else just return None.
for i in range(0, 10):
node = random.choice(self.nodes)
if self._is_node_in_multi_rack_dc(node) == True:
return node.get_connection()
return None