Prerequisites

Prerequisites for deploying TiKV

Typical deployments of TiKV include a number of components:

  • 3+ TiKV nodes
  • 3+ Placement Driver (PD) nodes
  • 1 Monitoring node
  • 1 or more client application or query layer (like TiDB)
TiKV is deployed alongside a Placement Driver (PD) cluster. PD is the cluster manager of TiKV, which periodically checks replication constraints to balance load and data automatically.

Your first steps into TiKV require only the following:

After you set up the environment, follow through the Try guide to get a test setup of TiKV running on your machine.

Production usage is typically done via automation requiring:

If you have your production environment ready, follow through the Ansible deployment guide. You may optionally choose unsupported manual Docker deployment or binary deployment strategies.

Finally, if you want to build your own binary TiKV you should consult the README of the repository.

System requirements

The minimum specifications for testing or developing TiKV or PD are:

  • 2+ core
  • 8+ GB RAM
  • An SSD

TiKV hosts must support the x86-64 architecture and the SSE 4.2 instruction set.

TiKV works well in VMWare, KVM, and Xen virtual machines.

Production Specifications

The suggested PD specifications for production are:

  • 3+ nodes
  • 4+ cores
  • 8+ GB RAM, with no swap space.
  • 200+ GB Optane, NVMe, or SSD drive
  • 10 Gigabit ethernet (2x preferred)
  • A Linux Operating System (PD is most widely tested on CentOS 7).

The suggested TiKV specifications for production are:

  • 3+ nodes
  • 16+ cores
  • 32+ GB RAM, with no swap space.
  • 200+ GB Optane, NVMe, or SSD drive (Under 1.5 TB capacity is ideal in our tests)
  • 10 Gigabit ethernet (2x preferred)
  • A Linux Operating System (TiKV is most widely tested on CentOS 7).

Network requirements

TiKV deployments require total connectivity of all services. Each TiKV, PD, and client must be able to reach each all other and advertise the addresses of all other services to new services. This connectivity allows TiKV and PD to replicate and balance data resiliently across the entire deployment.

If the hosts are not already able to reach each other, it is possible to accomplish this through a Virtual Local Area Network (VLAN). Speak to your system administrator to explore your options.

TiKV requires the following network port configuration to run. Based on the TiKV deployment in actual environments, the administrator can open relevant ports in the network side and host side.

ComponentDefault PortProtocolDescription
TiKV20160gRPCClient (such as Query Layers) port.
TiKV20180TextStatus port, Prometheus metrics at /metrics.
PD2379gRPCThe client port, for communication with clients.
PD2380gRPCThe server port, for communication with TiKV.
If you are deploying tools alongside TiKV you may need to open or configure other ports. For example, port 3000 for the Grafana service.

You can ensure your configuration is correct by creating echo servers on the ports/IPs by using ncat (from the nmap package):

ncat -l $PORT -k -c 'xargs -n1 echo'

Then from the other machines, verify that the echo server is reachable with curl $IP:$PORT.

Optional: Configure Monitoring

TiKV can work with Prometheus and Grafana to provide a rich visual monitoring dashboard. This comes preconfigured if you use the Ansible or Docker Compose deployment methods.

We strongly recommend using an up-to-date version of Mozilla Firefox or Google Chrome when accessing Grafana.