Introduction
The following describes the usage and examples of each instruction, where $IOTDB_CLUSTER_HOME indicates the path of the distributed IoTDB installation directory.
Prerequisites
To use the nodetool, you need to enable JMX service. Please refer to for details.
Instructions
The nodetool shell tool startup script is located at $IOTDB_CLUSTER_HOME/sbin folder, you can specify the IP address and port of the cluster during startup.
IP is the IP (or hostname) of the node that you expect to connect to, and port is the JMX service port specified when the IoTDB cluster is started.
The default values are and 31999
, respectively.
If you need to monitor the remote cluster or modify the JMX service port number, set the actual IP and port with the -h
and -p
options.
When JMX authentication service is enabled, the username and password of JMX service should be specified. The default values are iotdb
and passw!d
, respectively. Please use the start parameters -u
and -pw
to specify the username and password of the JMX service.
Explains
In a distributed system, a node is identified by node IP, metadata port, data port and cluster port <IP:METAPORT:DATAPORT:CLUSTERPORT>.
Distributed IoTDB uses consistent hash to support data distribution.
You can know each node in the cluster by command ring
, which prints node ring information.
1.Input
2.Output
The output is a multi-line string, and each line of string is a key value pair, where the key represents the node identifier, and the value represents the node (IP:METAPORT:DATAPORT:CLUSTERPORT), the format is <key -> value>.
3.Examples
Suppose that the current cluster runs on three nodes: 127.0.0.1:9003:40010:6667, 127.0.0.1:9005:40012:6668, and 127.0.0.1:9007:40014:6669.
Examples of input instructions for different systems are as follows:
Linux and MacOS:
Windows:
Shell > .\sbin\nodetool.bat -h 127.0.0.1 -p 31999 ring
Press enter to execute the command.
The output of the example instruction is as follows:
Node Identifier Node
330411070 -> 127.0.0.1:9003:40010:6667
330454032 -> 127.0.0.1:9005:40012:6668
The above output shows that there are three nodes in the current cluster, and the output results are ordered by their identifier ascendant.
The time series metadata of distributed iotdb is divided into multiple data groups according to their storage groups, in which the storage group and data partition are many to one relationship.
That is, all metadata of a storage group only exists in the same data group, and a data group may contain multiple storage groups.
The data partition is composed of several replica nodes to ensure high availability of data, and one of the nodes plays the role of leader.
Through this instruction, the user can know the metadata of a certain path, and the nodes under which the data is stored.
1.Input
The instruction for querying data partition information is
partition
. The parameters are described as follows:
2.Output
3.Examples
Suppose that the current cluster runs on three nodes: 127.0.0.1:9003:40010:6667, 127.0.0.1:9005:40012:6668, and 127.0.0.1:9007:40014:6669.
The number of copies is 2 and there are 3 storage groups:{ root.beijing , root.shanghai , root.guangzhou}.
Linux and MacOS:
Shell > ./sbin/nodetool.sh -h 127.0.0.1 -p 31999 partition -path root.guangzhou.d1
Windows:
Shell > .\sbin\nodetool.bat -h 127.0.0.1 -p 31999 partition -path root.guangzhou.d1
Press enter to execute the command.
The output of the example instruction is as follows:
DATA<root.guangzhou.d1, 1576723735188, 1576723735188> -> [127.0.0.1:9003:40010:6667, 127.0.0.1:9005:40012:6668]
- Partition of query data (specified time range, time partition interval is one day)
Linux and MacOS:
Shell > ./sbin/nodetool.sh -h 127.0.0.1 -p 31999 partition -path root.guangzhou.d1 -st 1576624778159 -et 1576724778159
Windows:
Press enter to execute the command.
The output of the example instruction is as follows:
DATA<root.guangzhou.d1, 1576627200000, 1576713599999> -> [127.0.0.1:9007:40014:6669, 127.0.0.1:9003:40010:6667]
DATA<root.guangzhou.d1, 1576713600000, 1576724778159> -> [127.0.0.1:9003:40010:6667, 127.0.0.1:9005:40012:6668]
DATA<root.guangzhou.d1, 1576624778159, 1576627199999> -> [127.0.0.1:9005:40012:6668, 127.0.0.1:9007:40014:6669]
- Query metadata partition
Linux and MacOS:
Shell > ./sbin/nodetool.sh -h 127.0.0.1 -p 31999 partition -path root.guangzhou.d1 -m
Windows:
Shell > .\sbin\nodetool.bat -h 127.0.0.1 -p 31999 partition -path root.guangzhou.d1 -m
Press enter to execute the command.
The output of the example instruction is as follows:
DATA<root.guangzhou.d1, 1576723735188, 1576723735188> -> [127.0.0.1:9003:40010, 127.0.0.1:9004:40011]
The above output shows that the data group to which root.guangzhou.d1 belongs contains two nodes, of which 127.0.0.1:9003:40010 is the header node.
Distributed IoTDB divides data into a fixed number of (10000 by default) slots, and the leader of the cluster management group divides the slots among data groups.
Through this instruction, you can know the number of slots managed by each data group.
- Input
The parameters are described as follows:
Parameter | Description | Examples |
---|---|---|
-a or —all | Query the number of slots managed by all data groups. By default only data groups of the query node are shown | -a |
2.Output
The output is a multi-line string, in which each line is a key-value pair, where the key represents the data group, and the value represents the number of slots managed, and the format is <key -> value>.
3.Examples
Linux and MacOS:
Windows:
Shell > .\sbin\nodetool.bat -h 127.0.0.1 -p 31999 host
Press enter to execute the command.
The output of the example instruction is as follows:
- All Partition Groups
Linux and MacOS:
Shell > ./sbin/nodetool.sh -h 127.0.0.1 -p 31999 host -a
Windows:
Shell > .\sbin\nodetool.bat -h 127.0.0.1 -p 31999 host -a
Press enter to execute the command.
The output of the example instruction is as follows:
Raft group Slot Number
(127.0.0.1:9003:40010:6667, 127.0.0.1:9005:40012:6668) -> 3333
(127.0.0.1:9005:40012:6668, 127.0.0.1:9007:40014:6669) -> 3333
(127.0.0.1:9007:40014:6669, 127.0.0.1:9003:40010:6667) -> 3334
Distributed IoTDB contains multiple nodes. For any node, there is a possibility that it cannot provide services normally due to network or hardware problems.
Through this instruction, you can know the current status of all nodes in the cluster.
1.Input
status
2.Output
The output is a multi-line string, where each line is a key-value pair, where the key represents the node (IP: METAPORT:DATAPORT), the value indicates the status of the node, “on” is normal, “off” is abnormal, and the format is < key -> value>.
3.Examples
Suppose that the current cluster runs on three nodes: 127.0.0.1:9003:40010:6667, 127.0.0.1:9005:40012:6668, and 127.0.0.1:9007:40014:6669, and the number of copies is 2.
Linux and MacOS:
Shell > ./sbin/nodetool.sh -h 127.0.0.1 -p 31999 status
Windows:
Shell > .\sbin\nodetool.bat -h 127.0.0.1 -p 31999 status
Press enter to execute the command.
Node Status
127.0.0.1:9003:40010:6667 -> on
The above output indicates that 127.0.0.1:9003:40010:6667 nodes and 127.0.0.1:9007:40014:6669 nodes are in normal state, and 127.0.0.1:9005:40012:6668 nodes cannot provide services.