Apr 11, 2013

How to add new node to an existing Aerospike Cluster - Using multicast

Modify /etc/citrusleaf/citrusleaf.conf as my example.
network {
service {
address any
port 3000
reuse-address
network-interface-name bond0x #If you use non standard interface name, Use this config.
#standard interface e.g. eth0, bond0, WLAN0.
}
heartbeat {
mode multicast
address 239.1.99.108
port 9918
interface-address 172.28.206.72 #You must set this config,
#if you wnat to add new node to an existing cluster.
#Each node must have unique ip address!!!
#Just set this config as node's ip addres.
interval 150
timeout 15
}
fabric {
address any
port 3001
}
info {
address any
port 3003
}
}
view raw citrusleaf.conf hosted with ❤ by GitHub

After modifying just start new node, it will be added an existing cluster automatically. You can confirm it on /var/log/citrusleaf.log with the following message.
Apr 11 2013 08:10:42 GMT: INFO (info): (base/thr_info.c:3078)  migrates in progress ( 1426 , 0 ) ::: ClusterSize 2 ::: objects 1
view raw gistfile1.md hosted with ❤ by GitHub


Tip(?).
I changed ip address to 172.28.206.73 - it's wrong address -,  after I got succeeded with above configuration. So I fixed it to 172.28.206.72 and restarted node. But I got some non normal message.
I issued the command and failed to get expected result - ClusterSize 2 -.
Apr 11 2013 05:17:23 GMT: INFO (paxos): (fabric/paxos.c:1948) Cluster Integrity Check: Detected succession list discrepancy between node bb9f5e6171cb190 and self bb935d8171cb190
Apr 11 2013 05:17:23 GMT: INFO (paxos): (fabric/paxos.c:1990) CLUSTER INTEGRITY FAULT. [Phase 1 of 2] To fix, issue this command across all nodes:  dun:nodes=bb9f5e6171cb190
Apr 11 2013 05:17:24 GMT: INFO (info): (base/thr_info.c:3071)  system memory: free 12087228kb ( 99 percent free )
Apr 11 2013 05:17:24 GMT: INFO (info): (base/thr_info.c:3078)  migrates in progress ( 0 , 0 ) ::: ClusterSize 1 ::: objects 0

$ clinfo -v 'dun:nodes=bb9f5e6171cb190'

Apr 11 2013 08:26:27 GMT: INFO (info): (base/thr_info.c:3078)  migrates in progress ( 0 , 0 ) ::: ClusterSize 1 ::: objects 1
view raw gistfile1.md hosted with ❤ by GitHub

I fixed this event by restarting all nodes.

No comments: