# Scale

## Learn how to scale out/in TiKV using TiUP

The capacity of a TiKV cluster can be increased or decreased without interrupting online services.

This document describes how to scale a TiKV or PD cluster using TiUP.

For example, assume that the topology of the cluster is as follows:

Host IPService
10.0.1.1Monitor
10.0.1.2PD
10.0.1.3PD
10.0.1.4PD
10.0.1.5TiKV
10.0.1.6TiKV
10.0.1.7TiKV

## Scale out a TiKV cluster

If you want to add a TiKV node to the 10.0.1.8 host, take the following steps.

1. Configure the scale-out topology

Put the following contents in the scale-out-tikv.yaml file:

tikv_servers:
- host: 10.0.1.8
ssh_port: 22
port: 20160
status_port: 20180
deploy_dir: /data/deploy/install/deploy/tikv-20160
data_dir: /data/deploy/install/data/tikv-20160
log_dir: /data/deploy/install/log/tikv-20160


To view the configuration of the current cluster, run tiup cluster edit-config <cluster-name>. Because the parameter configuration of global and server_configs is inherited by scale-out-tikv.yaml and thus also takes effect in scale-out-tikv.yaml.

2. Run the scale-out command

tiup cluster scale-out <cluster-name> scale-out-tikv.yaml


If you see the message “Scaled cluster out successfully”, it means that the scale-out operation is successfully completed.

3. Check the cluster status

tiup cluster display <cluster-name>


Access the monitoring platform at http://10.0.1.1:3000 using your browser to monitor the status of the cluster and the new node.

After the scale-out, the cluster topology is as follows:

Host IPService
10.0.1.1Monitor
10.0.1.2PD
10.0.1.3PD
10.0.1.4PD
10.0.1.5TiKV
10.0.1.6TiKV
10.0.1.7TiKV
10.0.1.8TiKV

## Scale out a PD cluster

If you want to add a PD node to the 10.0.1.9 host, take the following steps.

1. Configure the scale-out topology

Put the following contents in the scale-out-pd.yaml file:

pd_servers:
- host: 10.0.1.9
ssh_port: 22
client_port: 2379
peer_port: 2380
deploy_dir: /data/deploy/install/deploy/pd-2379
data_dir: /data/deploy/install/data/pd-2379
log_dir: /data/deploy/install/log/pd-2379


To view the configuration of the current cluster, run tiup cluster edit-config <cluster-name>. Because the parameter configuration of global and server_configs is inherited by scale-out-pd.yaml and thus also takes effect in scale-out-pd.yaml.

2. Run the scale-out command

tiup cluster scale-out <cluster-name> scale-out-pd.yaml


If you see the message “Scaled cluster out successfully”, it means that the scale-out operation is successfully completed.

3. Check the cluster status

tiup cluster display <cluster-name>


Access the monitoring platform at http://10.0.1.1:3000 using your browser to monitor the status of the cluster and the new node.

After the scale-out, the cluster topology is as follows:

Host IPService
10.0.1.1Monitor
10.0.1.2PD
10.0.1.3PD
10.0.1.4PD
10.0.1.5TiKV
10.0.1.6TiKV
10.0.1.7TiKV
10.0.1.8TiKV
10.0.1.9PD

## Scale in a TiKV cluster

If you want to remove a TiKV node from the 10.0.1.5 host, take the following steps.

You can take similar steps to remove a PD node.
1. View the node ID information:

tiup cluster display <cluster-name>

2. Run the scale-in command:

tiup cluster scale-in <cluster-name> --node 10.0.1.5:20160


The --node parameter is the ID of the node to be taken offline.

If you see the message “Scaled cluster in successfully”, it means that the scale-in operation is successfully completed.

Besides, if the status of the node to be taken offline becomes Tombstone, it also indicates that the scale-in operation is successfully completed because the scale-in process takes some time.

3. Check the cluster status:

To check the scale-in status, run the following command:

tiup cluster display <cluster-name>


Access the monitoring platform at http://10.0.1.1:3000 using your browser, and view the status of the cluster.

After the scale-in, the current topology is as follows:

Host IPService
10.0.1.1Monitor
10.0.1.2PD
10.0.1.3PD
10.0.1.4PD
10.0.1.6TiKV
10.0.1.7TiKV
10.0.1.8TiKV
10.0.1.9PD