From 8a043303c358498629346143318de68334db17b4 Mon Sep 17 00:00:00 2001 From: Xuecheng Zhang Date: Fri, 20 Oct 2023 17:33:30 +0800 Subject: [PATCH] en,zh: bump operator to v1.5.1 for release-1.5 (#2460) --- en/TOC.md | 2 +- en/_index.md | 2 +- en/access-dashboard.md | 2 +- en/advanced-statefulset.md | 4 +- en/aggregate-multiple-cluster-monitor-data.md | 6 +- en/backup-restore-cr.md | 4 +- en/backup-to-aws-s3-by-snapshot.md | 2 +- en/backup-to-aws-s3-using-br.md | 2 +- en/backup-to-azblob-using-br.md | 2 +- en/backup-to-gcs-using-br.md | 2 +- en/backup-to-gcs.md | 2 +- en/backup-to-pv-using-br.md | 2 +- en/backup-to-s3.md | 4 +- en/cheat-sheet.md | 6 +- en/configure-a-tidb-cluster.md | 2 +- en/configure-storage-class.md | 4 +- en/deploy-heterogeneous-tidb-cluster.md | 2 +- en/deploy-on-alibaba-cloud.md | 6 +- en/deploy-on-aws-eks.md | 10 ++-- en/deploy-on-azure-aks.md | 10 ++-- en/deploy-on-gcp-gke.md | 10 ++-- ...tidb-cluster-across-multiple-kubernetes.md | 2 +- en/deploy-tidb-dm.md | 2 +- en/deploy-tidb-from-kubernetes-gke.md | 8 +-- ...tidb-monitor-across-multiple-kubernetes.md | 4 +- en/deploy-tidb-operator.md | 28 ++++----- en/enable-monitor-dynamic-configuration.md | 2 +- en/enable-monitor-shards.md | 2 +- en/get-started.md | 10 ++-- en/initialize-a-cluster.md | 2 +- en/monitor-a-tidb-cluster.md | 6 +- en/restore-from-aws-s3-by-snapshot.md | 2 +- en/restore-from-aws-s3-using-br.md | 4 +- en/restore-from-azblob-using-br.md | 4 +- en/restore-from-gcs-using-br.md | 4 +- en/restore-from-gcs.md | 2 +- en/restore-from-pv-using-br.md | 2 +- en/restore-from-s3.md | 2 +- en/tidb-toolkit.md | 20 +++---- en/upgrade-tidb-operator.md | 60 +++++++++---------- zh/TOC.md | 2 +- zh/_index.md | 2 +- zh/access-dashboard.md | 2 +- zh/advanced-statefulset.md | 4 +- zh/aggregate-multiple-cluster-monitor-data.md | 6 +- zh/backup-restore-cr.md | 4 +- zh/backup-to-aws-s3-by-snapshot.md | 2 +- zh/backup-to-aws-s3-using-br.md | 2 +- zh/backup-to-azblob-using-br.md | 2 +- zh/backup-to-gcs-using-br.md | 2 +- zh/backup-to-gcs.md | 2 +- zh/backup-to-pv-using-br.md | 2 +- zh/backup-to-s3.md | 4 +- zh/cheat-sheet.md | 6 +- zh/configure-a-tidb-cluster.md | 2 +- zh/configure-storage-class.md | 2 +- zh/deploy-heterogeneous-tidb-cluster.md | 2 +- zh/deploy-on-alibaba-cloud.md | 6 +- zh/deploy-on-aws-eks.md | 10 ++-- zh/deploy-on-azure-aks.md | 10 ++-- zh/deploy-on-gcp-gke.md | 10 ++-- ...tidb-cluster-across-multiple-kubernetes.md | 2 +- zh/deploy-tidb-dm.md | 2 +- zh/deploy-tidb-from-kubernetes-gke.md | 8 +-- ...tidb-monitor-across-multiple-kubernetes.md | 4 +- zh/deploy-tidb-operator.md | 28 ++++----- zh/enable-monitor-dynamic-configuration.md | 2 +- zh/enable-monitor-shards.md | 2 +- zh/get-started.md | 22 +++---- zh/initialize-a-cluster.md | 2 +- zh/monitor-a-tidb-cluster.md | 8 +-- zh/restore-from-aws-s3-by-snapshot.md | 2 +- zh/restore-from-aws-s3-using-br.md | 4 +- zh/restore-from-azblob-using-br.md | 4 +- zh/restore-from-gcs-using-br.md | 4 +- zh/restore-from-gcs.md | 2 +- zh/restore-from-pv-using-br.md | 2 +- zh/restore-from-s3.md | 2 +- zh/tidb-toolkit.md | 20 +++---- zh/upgrade-tidb-operator.md | 60 +++++++++---------- 80 files changed, 258 insertions(+), 258 deletions(-) diff --git a/en/TOC.md b/en/TOC.md index 55c6954b7..b92eb5abd 100644 --- a/en/TOC.md +++ b/en/TOC.md @@ -113,7 +113,7 @@ - [Advanced StatefulSet Controller](advanced-statefulset.md) - [Admission Controller](enable-admission-webhook.md) - [Sysbench Performance Test](benchmark-sysbench.md) - - [API References](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md) + - [API References](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md) - [Cheat Sheet](cheat-sheet.md) - [Required RBAC Rules](tidb-operator-rbac.md) - Tools diff --git a/en/_index.md b/en/_index.md index 9c459a75b..479d24f6f 100644 --- a/en/_index.md +++ b/en/_index.md @@ -70,7 +70,7 @@ hide_commit: true -[API Docs](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md) +[API Docs](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md) [Tools](https://docs.pingcap.com/tidb-in-kubernetes/dev/tidb-toolkit) diff --git a/en/access-dashboard.md b/en/access-dashboard.md index 02f634e24..8ec290621 100644 --- a/en/access-dashboard.md +++ b/en/access-dashboard.md @@ -244,7 +244,7 @@ To enable this feature, you need to deploy TidbNGMonitoring CR using TiDB Operat EOF ``` - For more configuration items of the TidbNGMonitoring CR, see [example in tidb-operator](https://github.com/pingcap/tidb-operator/blob/v1.5.0/examples/advanced/tidb-ng-monitoring.yaml). + For more configuration items of the TidbNGMonitoring CR, see [example in tidb-operator](https://github.com/pingcap/tidb-operator/blob/v1.5.1/examples/advanced/tidb-ng-monitoring.yaml). 3. Enable Continuous Profiling. diff --git a/en/advanced-statefulset.md b/en/advanced-statefulset.md index d89d2b7e4..e429802b4 100644 --- a/en/advanced-statefulset.md +++ b/en/advanced-statefulset.md @@ -21,7 +21,7 @@ The [advanced StatefulSet controller](https://github.com/pingcap/advanced-statef {{< copyable "shell-regular" >}} ```shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/advanced-statefulset-crd.v1beta1.yaml + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/advanced-statefulset-crd.v1beta1.yaml ``` * For Kubernetes versions >= 1.16: @@ -29,7 +29,7 @@ The [advanced StatefulSet controller](https://github.com/pingcap/advanced-statef {{< copyable "shell-regular" >}} ``` - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/advanced-statefulset-crd.v1.yaml + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/advanced-statefulset-crd.v1.yaml ``` 2. Enable the `AdvancedStatefulSet` feature in `values.yaml` of the TiDB Operator chart: diff --git a/en/aggregate-multiple-cluster-monitor-data.md b/en/aggregate-multiple-cluster-monitor-data.md index 61b53ae32..0dac100d8 100644 --- a/en/aggregate-multiple-cluster-monitor-data.md +++ b/en/aggregate-multiple-cluster-monitor-data.md @@ -24,7 +24,7 @@ Thanos provides [Thanos Query](https://thanos.io/tip/components/query.md/) compo {{< copyable "shell-regular" >}} ```shell - kubectl -n ${namespace} apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/monitor-with-thanos/tidb-monitor.yaml + kubectl -n ${namespace} apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/monitor-with-thanos/tidb-monitor.yaml ``` 2. Deploy the Thanos Query component. @@ -34,7 +34,7 @@ Thanos provides [Thanos Query](https://thanos.io/tip/components/query.md/) compo {{< copyable "shell-regular" >}} ``` - curl -sl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/monitor-with-thanos/thanos-query.yaml + curl -sl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/monitor-with-thanos/thanos-query.yaml ``` 2. Manually modify the `--store` parameter in the `thanos-query.yaml` file by updating `basic-prometheus:10901` to `basic-prometheus.${namespace}:10901`. @@ -182,4 +182,4 @@ spec: After RemoteWrite is enabled, Prometheus pushes the monitoring data to [Thanos Receiver](https://thanos.io/tip/components/receive.md/). For more information, refer to [the design of Thanos Receiver](https://thanos.io/v0.8/proposals/201812_thanos-remote-receive/). -For details on the deployment, refer to [this example of integrating TidbMonitor with Thanos Receiver](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-prom-remotewrite). +For details on the deployment, refer to [this example of integrating TidbMonitor with Thanos Receiver](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-prom-remotewrite). diff --git a/en/backup-restore-cr.md b/en/backup-restore-cr.md index 34512610e..fe47d61d9 100644 --- a/en/backup-restore-cr.md +++ b/en/backup-restore-cr.md @@ -24,7 +24,7 @@ This section introduces the fields in the `Backup` CR. - If an image is specified without the version, such as `.spec.toolImage: private/registry/br`, the `private/registry/br:${tikv_version}` image is used for backup. - When using Dumpling for backup, you can specify the Dumpling version in this field. - If the Dumpling version is specified in this field, such as `spec.toolImage: pingcap/dumpling:v7.1.1`, the image of the specified version is used for backup. - - If the field is not specified, the Dumpling version specified in `TOOLKIT_VERSION` of the [Backup Manager Dockerfile](https://github.com/pingcap/tidb-operator/blob/v1.5.0/images/tidb-backup-manager/Dockerfile) is used for backup by default. + - If the field is not specified, the Dumpling version specified in `TOOLKIT_VERSION` of the [Backup Manager Dockerfile](https://github.com/pingcap/tidb-operator/blob/v1.5.1/images/tidb-backup-manager/Dockerfile) is used for backup by default. * `.spec.backupType`: the backup type. This field is valid only when you use BR for backup. Currently, the following three types are supported, and this field can be combined with the `.spec.tableFilter` field to configure table filter rules: * `full`: back up all databases in a TiDB cluster. @@ -261,7 +261,7 @@ This section introduces the fields in the `Restore` CR. * `.spec.toolImage`:the tools image used by `Restore`. TiDB Operator supports this configuration starting from v1.1.9. - When using BR for restoring, you can specify the BR version in this field. For example,`spec.toolImage: pingcap/br:v7.1.1`. If not specified, `pingcap/br:${tikv_version}` is used for restoring by default. - - When using Lightning for restoring, you can specify the Lightning version in this field. For example, `spec.toolImage: pingcap/lightning:v7.1.1`. If not specified, the Lightning version specified in `TOOLKIT_VERSION` of the [Backup Manager Dockerfile](https://github.com/pingcap/tidb-operator/blob/v1.5.0/images/tidb-backup-manager/Dockerfile) is used for restoring by default. + - When using Lightning for restoring, you can specify the Lightning version in this field. For example, `spec.toolImage: pingcap/lightning:v7.1.1`. If not specified, the Lightning version specified in `TOOLKIT_VERSION` of the [Backup Manager Dockerfile](https://github.com/pingcap/tidb-operator/blob/v1.5.1/images/tidb-backup-manager/Dockerfile) is used for restoring by default. * `.spec.backupType`: the restore type. This field is valid only when you use BR to restore data. Currently, the following three types are supported, and this field can be combined with the `.spec.tableFilter` field to configure table filter rules: * `full`: restore all databases in a TiDB cluster. diff --git a/en/backup-to-aws-s3-by-snapshot.md b/en/backup-to-aws-s3-by-snapshot.md index 8d660d9ff..8a76f5e07 100644 --- a/en/backup-to-aws-s3-by-snapshot.md +++ b/en/backup-to-aws-s3-by-snapshot.md @@ -42,7 +42,7 @@ The following sections exemplify how to back up data of the TiDB cluster `demo1` ### Step 1. Set up the environment for EBS volume snapshot backup -1. Download the file [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml) to the backup server. +1. Download the file [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml) to the backup server. 2. Create the RBAC-related resources required for the backup in the `test1` namespace by running the following command: diff --git a/en/backup-to-aws-s3-using-br.md b/en/backup-to-aws-s3-using-br.md index 802c62efa..9a69cf642 100644 --- a/en/backup-to-aws-s3-using-br.md +++ b/en/backup-to-aws-s3-using-br.md @@ -51,7 +51,7 @@ This document provides an example about how to back up the data of the `demo1` T kubectl create namespace backup-test ``` -2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `backup-test` namespace: +2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `backup-test` namespace: ```shell kubectl apply -f backup-rbac.yaml -n backup-test diff --git a/en/backup-to-azblob-using-br.md b/en/backup-to-azblob-using-br.md index 3c158c5dd..46038d102 100644 --- a/en/backup-to-azblob-using-br.md +++ b/en/backup-to-azblob-using-br.md @@ -48,7 +48,7 @@ This document provides an example about how to back up the data of the `demo1` T kubectl create namespace backup-test ``` -2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `backup-test` namespace: +2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `backup-test` namespace: ```shell kubectl apply -f backup-rbac.yaml -n backup-test diff --git a/en/backup-to-gcs-using-br.md b/en/backup-to-gcs-using-br.md index 76d6abd7e..bd70eaf12 100644 --- a/en/backup-to-gcs-using-br.md +++ b/en/backup-to-gcs-using-br.md @@ -48,7 +48,7 @@ This document provides an example about how to back up the data of the `demo1` T kubectl create namespace backup-test ``` -2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `test1` namespace: +2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `test1` namespace: ```shell kubectl apply -f backup-rbac.yaml -n backup-test diff --git a/en/backup-to-gcs.md b/en/backup-to-gcs.md index 24a19f9fc..9bf53cc02 100644 --- a/en/backup-to-gcs.md +++ b/en/backup-to-gcs.md @@ -38,7 +38,7 @@ To better explain how to perform the backup operation, this document shows an ex ### Step 1: Prepare for ad-hoc full backup -1. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml) and execute the following command to create the role-based access control (RBAC) resources in the `test1` namespace: +1. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml) and execute the following command to create the role-based access control (RBAC) resources in the `test1` namespace: {{< copyable "shell-regular" >}} diff --git a/en/backup-to-pv-using-br.md b/en/backup-to-pv-using-br.md index fcdc71034..e6ecdb66f 100644 --- a/en/backup-to-pv-using-br.md +++ b/en/backup-to-pv-using-br.md @@ -33,7 +33,7 @@ This document provides an example about how to back up the data of the `demo1` T ### Step 1: Prepare for an ad-hoc backup -1. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml) to the server that runs the backup task. +1. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml) to the server that runs the backup task. 2. Execute the following command to create the role-based access control (RBAC) resources in the `test1` namespace: diff --git a/en/backup-to-s3.md b/en/backup-to-s3.md index d21d89542..ced0df3ff 100644 --- a/en/backup-to-s3.md +++ b/en/backup-to-s3.md @@ -49,12 +49,12 @@ GRANT ### Step 1: Prepare for ad-hoc full backup -1. Execute the following command to create the role-based access control (RBAC) resources in the `tidb-cluster` namespace based on [backup-rbac.yaml](https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/backup/backup-rbac.yaml): +1. Execute the following command to create the role-based access control (RBAC) resources in the `tidb-cluster` namespace based on [backup-rbac.yaml](https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/backup/backup-rbac.yaml): {{< copyable "shell-regular" >}} ```shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/backup/backup-rbac.yaml -n tidb-cluster + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/backup/backup-rbac.yaml -n tidb-cluster ``` 2. Grant permissions to the remote storage. diff --git a/en/cheat-sheet.md b/en/cheat-sheet.md index d94a2ab82..0aecf7def 100644 --- a/en/cheat-sheet.md +++ b/en/cheat-sheet.md @@ -493,7 +493,7 @@ For example: {{< copyable "shell-regular" >}} ```shell -helm inspect values pingcap/tidb-operator --version=v1.5.0 > values-tidb-operator.yaml +helm inspect values pingcap/tidb-operator --version=v1.5.1 > values-tidb-operator.yaml ``` ### Deploy using Helm chart @@ -509,7 +509,7 @@ For example: {{< copyable "shell-regular" >}} ```shell -helm install tidb-operator pingcap/tidb-operator --namespace=tidb-admin --version=v1.5.0 -f values-tidb-operator.yaml +helm install tidb-operator pingcap/tidb-operator --namespace=tidb-admin --version=v1.5.1 -f values-tidb-operator.yaml ``` ### View the deployed Helm release @@ -533,7 +533,7 @@ For example: {{< copyable "shell-regular" >}} ```shell -helm upgrade tidb-operator pingcap/tidb-operator --version=v1.5.0 -f values-tidb-operator.yaml +helm upgrade tidb-operator pingcap/tidb-operator --version=v1.5.1 -f values-tidb-operator.yaml ``` ### Delete Helm release diff --git a/en/configure-a-tidb-cluster.md b/en/configure-a-tidb-cluster.md index 719a74c8e..ba076e023 100644 --- a/en/configure-a-tidb-cluster.md +++ b/en/configure-a-tidb-cluster.md @@ -24,7 +24,7 @@ If you are using a NUMA-based CPU, you need to enable `Static`'s CPU management ## Configure TiDB deployment -To configure a TiDB deployment, you need to configure the `TiDBCluster` CR. Refer to the [TidbCluster example](https://github.com/pingcap/tidb-operator/blob/v1.5.0/examples/advanced/tidb-cluster.yaml) for an example. For the complete configurations of `TiDBCluster` CR, refer to [API documentation](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md). +To configure a TiDB deployment, you need to configure the `TiDBCluster` CR. Refer to the [TidbCluster example](https://github.com/pingcap/tidb-operator/blob/v1.5.1/examples/advanced/tidb-cluster.yaml) for an example. For the complete configurations of `TiDBCluster` CR, refer to [API documentation](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md). > **Note:** > diff --git a/en/configure-storage-class.md b/en/configure-storage-class.md index b0d3d6c2e..3b3d6671f 100644 --- a/en/configure-storage-class.md +++ b/en/configure-storage-class.md @@ -95,7 +95,7 @@ The `/mnt/ssd`, `/mnt/sharedssd`, `/mnt/monitoring`, and `/mnt/backup` directori 1. Download the deployment file for the local-volume-provisioner. ```shell - wget https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/local-pv/local-volume-provisioner.yaml + wget https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/local-pv/local-volume-provisioner.yaml ``` 2. If you are using the same discovery directory as described in [Step 1: Pre-allocate local storage](#step-1-pre-allocate-local-storage), you can skip this step. If you are using a different path for the discovery directory than in the previous step, you need to modify the ConfigMap and DaemonSet spec. @@ -163,7 +163,7 @@ The `/mnt/ssd`, `/mnt/sharedssd`, `/mnt/monitoring`, and `/mnt/backup` directori 3. Deploy the `local-volume-provisioner`. ```shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/local-dind/local-volume-provisioner.yaml + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/local-dind/local-volume-provisioner.yaml ``` 4. Check the status of the Pod and PV. diff --git a/en/deploy-heterogeneous-tidb-cluster.md b/en/deploy-heterogeneous-tidb-cluster.md index 458f0adff..207de6cbb 100644 --- a/en/deploy-heterogeneous-tidb-cluster.md +++ b/en/deploy-heterogeneous-tidb-cluster.md @@ -165,7 +165,7 @@ After creating certificates, take the following steps to deploy a TLS-enabled he In the configuration file, `spec.tlsCluster.enabled`controls whether to enable TLS between the components and `spec.tidb.tlsClient.enabled`controls whether to enable TLS for the MySQL client. - - For more configurations of a TLS-enabled heterogeneous cluster, see the ['heterogeneous-tls'](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/heterogeneous-tls) example. + - For more configurations of a TLS-enabled heterogeneous cluster, see the ['heterogeneous-tls'](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/heterogeneous-tls) example. - For more configurations and field meanings of a TiDB cluster, see the [TiDB cluster configuration document](configure-a-tidb-cluster.md). 2. In the configuration file of your heterogeneous cluster, modify the configurations of each node according to your need. diff --git a/en/deploy-on-alibaba-cloud.md b/en/deploy-on-alibaba-cloud.md index 20520f8d7..7dc8d8836 100644 --- a/en/deploy-on-alibaba-cloud.md +++ b/en/deploy-on-alibaba-cloud.md @@ -89,7 +89,7 @@ All the instances except ACK mandatory workers are deployed across availability tikv_count = 3 tidb_count = 2 pd_count = 3 - operator_version = "v1.5.0" + operator_version = "v1.5.1" ``` * To deploy TiFlash in the cluster, set `create_tiflash_node_pool = true` in `terraform.tfvars`. You can also configure the node count and instance type of the TiFlash node pool by modifying `tiflash_count` and `tiflash_instance_type`. By default, the value of `tiflash_count` is `2`, and the value of `tiflash_instance_type` is `ecs.i2.2xlarge`. @@ -173,7 +173,7 @@ All the instances except ACK mandatory workers are deployed across availability cp manifests/dashboard.yaml.example tidb-dashboard.yaml ``` - To complete the CR file configuration, refer to [TiDB Operator API documentation](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md) and [Configure a TiDB Cluster](configure-a-tidb-cluster.md). + To complete the CR file configuration, refer to [TiDB Operator API documentation](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md) and [Configure a TiDB Cluster](configure-a-tidb-cluster.md). * To deploy TiFlash, configure `spec.tiflash` in `db.yaml` as follows: @@ -347,7 +347,7 @@ In the default configuration, the Terraform script creates a new VPC. To use the ### Configure the TiDB cluster -See [TiDB Operator API Documentation](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md) and [Configure a TiDB Cluster](configure-a-tidb-cluster.md). +See [TiDB Operator API Documentation](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md) and [Configure a TiDB Cluster](configure-a-tidb-cluster.md). ## Manage multiple TiDB clusters diff --git a/en/deploy-on-aws-eks.md b/en/deploy-on-aws-eks.md index 0ea2aa4f7..550b8f2dd 100644 --- a/en/deploy-on-aws-eks.md +++ b/en/deploy-on-aws-eks.md @@ -306,7 +306,7 @@ The following `c5d.4xlarge` example shows how to configure StorageClass for the 2. [Mount the local storage](https://github.com/kubernetes-sigs/sig-storage-local-static-provisioner/blob/master/docs/operations.md#use-a-whole-disk-as-a-filesystem-pv) to the `/mnt/ssd` directory. - 3. According to the mounting configuration, modify the [local-volume-provisioner.yaml](https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/eks/local-volume-provisioner.yaml) file. + 3. According to the mounting configuration, modify the [local-volume-provisioner.yaml](https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/eks/local-volume-provisioner.yaml) file. 4. Deploy and create a `local-storage` storage class using the modified `local-volume-provisioner.yaml` file. @@ -351,9 +351,9 @@ First, download the sample `TidbCluster` and `TidbMonitor` configuration files: {{< copyable "shell-regular" >}} ```shell -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aws/tidb-cluster.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aws/tidb-monitor.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aws/tidb-dashboard.yaml +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aws/tidb-cluster.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aws/tidb-monitor.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aws/tidb-dashboard.yaml ``` Refer to [configure the TiDB cluster](configure-a-tidb-cluster.md) to further customize and configure the CR before applying. @@ -668,4 +668,4 @@ Depending on the EKS cluster status, use different commands: Finally, execute `kubectl -n tidb-cluster apply -f tidb-cluster.yaml` to update the TiDB cluster configuration. -For detailed CR configuration, refer to [API references](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md) and [Configure a TiDB Cluster](configure-a-tidb-cluster.md). +For detailed CR configuration, refer to [API references](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md) and [Configure a TiDB Cluster](configure-a-tidb-cluster.md). diff --git a/en/deploy-on-azure-aks.md b/en/deploy-on-azure-aks.md index 733cacdd7..47ac3c01c 100644 --- a/en/deploy-on-azure-aks.md +++ b/en/deploy-on-azure-aks.md @@ -237,9 +237,9 @@ First, download the sample `TidbCluster` and `TidbMonitor` configuration files: {{< copyable "shell-regular" >}} ```shell -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aks/tidb-cluster.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aks/tidb-monitor.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aks/tidb-dashboard.yaml +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aks/tidb-cluster.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aks/tidb-monitor.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aks/tidb-dashboard.yaml ``` Refer to [configure the TiDB cluster](configure-a-tidb-cluster.md) to further customize and configure the CR before applying. @@ -526,7 +526,7 @@ Add a node pool for TiFlash/TiCDC respectively. You can set `--node-count` as re Finally, run the `kubectl -n tidb-cluster apply -f tidb-cluster.yaml` command to update the TiDB cluster configuration. -For detailed CR configuration, refer to [API references](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md) and [Configure a TiDB Cluster](configure-a-tidb-cluster.md). +For detailed CR configuration, refer to [API references](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md) and [Configure a TiDB Cluster](configure-a-tidb-cluster.md). ## Use other Disk volume types @@ -606,7 +606,7 @@ For instance types that provide local disks, refer to [Lsv2-series](https://docs {{< copyable "shell-regular" >}} ```shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/eks/local-volume-provisioner.yaml + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/eks/local-volume-provisioner.yaml ``` 3. Use local storage. diff --git a/en/deploy-on-gcp-gke.md b/en/deploy-on-gcp-gke.md index 5fc93765e..deccf0589 100644 --- a/en/deploy-on-gcp-gke.md +++ b/en/deploy-on-gcp-gke.md @@ -135,7 +135,7 @@ If you need to simulate bare-metal performance, some Google Cloud instance types {{< copyable "shell-regular" >}} ```shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/gke/local-ssd-provision/local-ssd-provision.yaml + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/gke/local-ssd-provision/local-ssd-provision.yaml ``` 3. Use the local storage. @@ -173,9 +173,9 @@ First, download the sample `TidbCluster` and `TidbMonitor` configuration files: {{< copyable "shell-regular" >}} ```shell -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/gcp/tidb-cluster.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/gcp/tidb-monitor.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/gcp/tidb-dashboard.yaml +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/gcp/tidb-cluster.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/gcp/tidb-monitor.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/gcp/tidb-dashboard.yaml ``` Refer to [configure the TiDB cluster](configure-a-tidb-cluster.md) to further customize and configure the CR before applying. @@ -467,4 +467,4 @@ The two components are *not required* in the deployment. This section shows a qu Finally, execute `kubectl -n tidb-cluster apply -f tidb-cluster.yaml` to update the TiDB cluster configuration. -For detailed CR configuration, refer to [API references](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md) and [Configure a TiDB Cluster](configure-a-tidb-cluster.md). +For detailed CR configuration, refer to [API references](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md) and [Configure a TiDB Cluster](configure-a-tidb-cluster.md). diff --git a/en/deploy-tidb-cluster-across-multiple-kubernetes.md b/en/deploy-tidb-cluster-across-multiple-kubernetes.md index ef1e0d029..087b07aed 100644 --- a/en/deploy-tidb-cluster-across-multiple-kubernetes.md +++ b/en/deploy-tidb-cluster-across-multiple-kubernetes.md @@ -642,7 +642,7 @@ If each Kubernetes have different Cluster Domain, you need to update the `spec.c After completing the above steps, this TidbCluster can be used as the initial TidbCluster for TiDB cluster deployment across Kubernetes clusters. You can refer the [section](#step-2-deploy-the-new-tidbcluster-to-join-the-tidb-cluster) to deploy other TidbCluster. -For more examples and development information, refer to [`multi-cluster`](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/multi-cluster). +For more examples and development information, refer to [`multi-cluster`](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/multi-cluster). ## Deploy TiDB monitoring components diff --git a/en/deploy-tidb-dm.md b/en/deploy-tidb-dm.md index eb43f2fe3..e6625fe17 100644 --- a/en/deploy-tidb-dm.md +++ b/en/deploy-tidb-dm.md @@ -17,7 +17,7 @@ summary: Learn how to deploy TiDB DM cluster on Kubernetes. ## Configure DM deployment -To configure the DM deployment, you need to configure the `DMCluster` Custom Resource (CR). For the complete configurations of the `DMCluster` CR, refer to the [DMCluster example](https://github.com/pingcap/tidb-operator/blob/v1.5.0/examples/dm/dm-cluster.yaml) and [API documentation](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md#dmcluster). Note that you need to choose the example and API of the current TiDB Operator version. +To configure the DM deployment, you need to configure the `DMCluster` Custom Resource (CR). For the complete configurations of the `DMCluster` CR, refer to the [DMCluster example](https://github.com/pingcap/tidb-operator/blob/v1.5.1/examples/dm/dm-cluster.yaml) and [API documentation](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md#dmcluster). Note that you need to choose the example and API of the current TiDB Operator version. ### Cluster name diff --git a/en/deploy-tidb-from-kubernetes-gke.md b/en/deploy-tidb-from-kubernetes-gke.md index c06db4ee8..e000720da 100644 --- a/en/deploy-tidb-from-kubernetes-gke.md +++ b/en/deploy-tidb-from-kubernetes-gke.md @@ -97,7 +97,7 @@ If you see `Ready` for all nodes, congratulations. You've set up your first Kube TiDB Operator uses [Custom Resource Definition (CRD)](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#customresourcedefinitions) to extend Kubernetes. Therefore, to use TiDB Operator, you must first create the `TidbCluster` CRD. ```shell -kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/crd.yaml && \ +kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/crd.yaml && \ kubectl get crd tidbclusters.pingcap.com ``` @@ -109,7 +109,7 @@ After the `TidbCluster` CRD is created, install TiDB Operator in your Kubernetes ```shell kubectl create namespace tidb-admin -helm install --namespace tidb-admin tidb-operator pingcap/tidb-operator --version v1.5.0 +helm install --namespace tidb-admin tidb-operator pingcap/tidb-operator --version v1.5.1 kubectl get po -n tidb-admin -l app.kubernetes.io/name=tidb-operator ``` @@ -126,13 +126,13 @@ To deploy the TiDB cluster, perform the following steps: 2. Deploy the TiDB cluster: ``` shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic/tidb-cluster.yaml -n demo + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic/tidb-cluster.yaml -n demo ``` 3. Deploy the TiDB cluster monitor: ``` shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic/tidb-monitor.yaml -n demo + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic/tidb-monitor.yaml -n demo ``` 4. View the Pod status: diff --git a/en/deploy-tidb-monitor-across-multiple-kubernetes.md b/en/deploy-tidb-monitor-across-multiple-kubernetes.md index 7a3ff4c04..8237e6696 100644 --- a/en/deploy-tidb-monitor-across-multiple-kubernetes.md +++ b/en/deploy-tidb-monitor-across-multiple-kubernetes.md @@ -24,7 +24,7 @@ The multiple Kubernetes clusters must meet the following condition: - The Prometheus (`TidbMonitor`) component in each Kubernetes cluster has access to the Thanos Receiver component. -For the deployment instructions of Thanos Receiver, refer to [kube-thanos](https://github.com/thanos-io/kube-thanos) and [the example](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-prom-remotewrite). +For the deployment instructions of Thanos Receiver, refer to [kube-thanos](https://github.com/thanos-io/kube-thanos) and [the example](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-prom-remotewrite). ### Architecture @@ -111,7 +111,7 @@ You need to configure the network and DNS of the Kubernetes clusters so that the - The Thanos Query component has access to the Pod IP of the Prometheus (`TidbMonitor`) component in each Kubernetes cluster. - The Thanos Query component has access to the Pod FQDN of the Prometheus (`TidbMonitor`) component in each Kubernetes cluster. -For the deployment instructions of Thanos Query, refer to [kube-thanos](https://github.com/thanos-io/kube-thanos) and [the example](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-with-thanos). +For the deployment instructions of Thanos Query, refer to [kube-thanos](https://github.com/thanos-io/kube-thanos) and [the example](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-with-thanos). #### Architecture diff --git a/en/deploy-tidb-operator.md b/en/deploy-tidb-operator.md index 1d22b29a7..3dc081062 100644 --- a/en/deploy-tidb-operator.md +++ b/en/deploy-tidb-operator.md @@ -45,7 +45,7 @@ TiDB Operator uses [Custom Resource Definition (CRD)](https://kubernetes.io/docs {{< copyable "shell-regular" >}} ```shell -kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/crd.yaml +kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/crd.yaml ``` If the server cannot access the Internet, you need to download the `crd.yaml` file on a machine with Internet access before installing: @@ -53,7 +53,7 @@ If the server cannot access the Internet, you need to download the `crd.yaml` fi {{< copyable "shell-regular" >}} ```shell -wget https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/crd.yaml +wget https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/crd.yaml kubectl create -f ./crd.yaml ``` @@ -101,7 +101,7 @@ When you use TiDB Operator, `tidb-scheduler` is not mandatory. Refer to [tidb-sc > **Note:** > - > `${chart_version}` represents the chart version of TiDB Operator. For example, `v1.5.0`. You can view the currently supported versions by running the `helm search repo -l tidb-operator` command. + > `${chart_version}` represents the chart version of TiDB Operator. For example, `v1.5.1`. You can view the currently supported versions by running the `helm search repo -l tidb-operator` command. 2. Configure TiDB Operator @@ -149,15 +149,15 @@ If your server cannot access the Internet, install TiDB Operator offline by the {{< copyable "shell-regular" >}} ```shell - wget http://charts.pingcap.org/tidb-operator-v1.5.0.tgz + wget http://charts.pingcap.org/tidb-operator-v1.5.1.tgz ``` - Copy the `tidb-operator-v1.5.0.tgz` file to the target server and extract it to the current directory: + Copy the `tidb-operator-v1.5.1.tgz` file to the target server and extract it to the current directory: {{< copyable "shell-regular" >}} ```shell - tar zxvf tidb-operator.v1.5.0.tgz + tar zxvf tidb-operator.v1.5.1.tgz ``` 2. Download the Docker images used by TiDB Operator @@ -169,8 +169,8 @@ If your server cannot access the Internet, install TiDB Operator offline by the {{< copyable "" >}} ```shell - pingcap/tidb-operator:v1.5.0 - pingcap/tidb-backup-manager:v1.5.0 + pingcap/tidb-operator:v1.5.1 + pingcap/tidb-backup-manager:v1.5.1 bitnami/kubectl:latest pingcap/advanced-statefulset:v0.3.3 k8s.gcr.io/kube-scheduler:v1.16.9 @@ -183,13 +183,13 @@ If your server cannot access the Internet, install TiDB Operator offline by the {{< copyable "shell-regular" >}} ```shell - docker pull pingcap/tidb-operator:v1.5.0 - docker pull pingcap/tidb-backup-manager:v1.5.0 + docker pull pingcap/tidb-operator:v1.5.1 + docker pull pingcap/tidb-backup-manager:v1.5.1 docker pull bitnami/kubectl:latest docker pull pingcap/advanced-statefulset:v0.3.3 - docker save -o tidb-operator-v1.5.0.tar pingcap/tidb-operator:v1.5.0 - docker save -o tidb-backup-manager-v1.5.0.tar pingcap/tidb-backup-manager:v1.5.0 + docker save -o tidb-operator-v1.5.1.tar pingcap/tidb-operator:v1.5.1 + docker save -o tidb-backup-manager-v1.5.1.tar pingcap/tidb-backup-manager:v1.5.1 docker save -o bitnami-kubectl.tar bitnami/kubectl:latest docker save -o advanced-statefulset-v0.3.3.tar pingcap/advanced-statefulset:v0.3.3 ``` @@ -199,8 +199,8 @@ If your server cannot access the Internet, install TiDB Operator offline by the {{< copyable "shell-regular" >}} ```shell - docker load -i tidb-operator-v1.5.0.tar - docker load -i tidb-backup-manager-v1.5.0.tar + docker load -i tidb-operator-v1.5.1.tar + docker load -i tidb-backup-manager-v1.5.1.tar docker load -i bitnami-kubectl.tar docker load -i advanced-statefulset-v0.3.3.tar ``` diff --git a/en/enable-monitor-dynamic-configuration.md b/en/enable-monitor-dynamic-configuration.md index f96042da9..f1415430e 100644 --- a/en/enable-monitor-dynamic-configuration.md +++ b/en/enable-monitor-dynamic-configuration.md @@ -38,7 +38,7 @@ spec: After you modify the `prometheusReloader` configuration, TidbMonitor restarts automatically. After the restart, the dynamic configuration feature is enabled. All configuration changes related to Prometheus are dynamically updated. -For more examples, refer to [monitor-dynamic-configmap](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-dynamic-configmap). +For more examples, refer to [monitor-dynamic-configmap](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-dynamic-configmap). ## Disable the dynamic configuration feature diff --git a/en/enable-monitor-shards.md b/en/enable-monitor-shards.md index 29d0036e1..c4231781b 100644 --- a/en/enable-monitor-shards.md +++ b/en/enable-monitor-shards.md @@ -49,4 +49,4 @@ spec: > - The number of Pods corresponding to TidbMonitor is the product of `replicas` and `shards`. For example, when `replicas` is `1` and `shards` is `2`, TiDB Operator creates 2 TidbMonitor Pods. > - After `shards` is changed, `Targets` are reallocated. However, the monitoring data already stored on the Pods is not reallocated. -For details on the configuration, refer to [shards example](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-shards). +For details on the configuration, refer to [shards example](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-shards). diff --git a/en/get-started.md b/en/get-started.md index d081adf94..d918235b4 100644 --- a/en/get-started.md +++ b/en/get-started.md @@ -175,7 +175,7 @@ First, you need to install the Custom Resource Definitions (CRDs) that are requi To install the CRDs, run the following command: ```shell -kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/crd.yaml +kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/crd.yaml ```
@@ -234,7 +234,7 @@ To install TiDB Operator, you can use [Helm 3](https://helm.sh/docs/intro/instal 3. Install TiDB Operator: ```shell - helm install --namespace tidb-admin tidb-operator pingcap/tidb-operator --version v1.5.0 + helm install --namespace tidb-admin tidb-operator pingcap/tidb-operator --version v1.5.1 ```
@@ -282,7 +282,7 @@ This section describes how to deploy a TiDB cluster and its monitoring services. ```shell kubectl create namespace tidb-cluster && \ - kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic/tidb-cluster.yaml + kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic/tidb-cluster.yaml ```
@@ -300,7 +300,7 @@ If you need to deploy a TiDB cluster on an ARM64 machine, refer to [Deploying a ### Deploy TiDB Dashboard independently ```shell -kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic/tidb-dashboard.yaml +kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic/tidb-dashboard.yaml ```
@@ -315,7 +315,7 @@ tidbdashboard.pingcap.com/basic created ### Deploy TiDB monitoring services ```shell -kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic/tidb-monitor.yaml +kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic/tidb-monitor.yaml ```
diff --git a/en/initialize-a-cluster.md b/en/initialize-a-cluster.md index a48e81e16..d9f5a01be 100644 --- a/en/initialize-a-cluster.md +++ b/en/initialize-a-cluster.md @@ -15,7 +15,7 @@ This document describes how to initialize a TiDB cluster on Kubernetes (K8s), sp ## Configure TidbInitializer -Refer to [TidbInitializer configuration example](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/initializer/tidb-initializer.yaml), [API documentation](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md), and the following steps to complete TidbInitializer Custom Resource (CR), and save it to the `${cluster_name}/tidb-initializer.yaml` file. When referring to the TidbInitializer configuration example and API documentation, you need to switch the branch to the TiDB Operator version currently in use. +Refer to [TidbInitializer configuration example](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/initializer/tidb-initializer.yaml), [API documentation](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md), and the following steps to complete TidbInitializer Custom Resource (CR), and save it to the `${cluster_name}/tidb-initializer.yaml` file. When referring to the TidbInitializer configuration example and API documentation, you need to switch the branch to the TiDB Operator version currently in use. ### Set the cluster namespace and name diff --git a/en/monitor-a-tidb-cluster.md b/en/monitor-a-tidb-cluster.md index 75ec74cb4..e06726da8 100644 --- a/en/monitor-a-tidb-cluster.md +++ b/en/monitor-a-tidb-cluster.md @@ -84,13 +84,13 @@ You can customize the Prometheus configuration by using a customized configurati 2. Set `spec.prometheus.config.configMapRef.name` and `spec.prometheus.config.configMapRef.namespace` to the name and namespace of the customized ConfigMap respectively. 3. Check if TidbMonitor has enabled [dynamic configuration](enable-monitor-dynamic-configuration.md). If not, you need to restart TidbMonitor's pod to reload the configuration. -For the complete configuration, refer to the [tidb-operator example](https://github.com/pingcap/tidb-operator/blob/v1.5.0/examples/monitor-with-externalConfigMap/prometheus/README.md). +For the complete configuration, refer to the [tidb-operator example](https://github.com/pingcap/tidb-operator/blob/v1.5.1/examples/monitor-with-externalConfigMap/prometheus/README.md). #### Add extra options to the command To add extra options to the command that starts Prometheus, configure `spec.prometheus.config.commandOptions`. -For the complete configuration, refer to the [tidb-operator example](https://github.com/pingcap/tidb-operator/blob/v1.5.0/examples/monitor-with-externalConfigMap/prometheus/README.md). +For the complete configuration, refer to the [tidb-operator example](https://github.com/pingcap/tidb-operator/blob/v1.5.1/examples/monitor-with-externalConfigMap/prometheus/README.md). > **Note:** > @@ -367,7 +367,7 @@ spec: imagePullPolicy: IfNotPresent ``` -For a complete configuration example, refer to [Example](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-multiple-cluster-non-tls) in the TiDB Operator repository. +For a complete configuration example, refer to [Example](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-multiple-cluster-non-tls) in the TiDB Operator repository. ### Monitor multiple clusters using Grafana diff --git a/en/restore-from-aws-s3-by-snapshot.md b/en/restore-from-aws-s3-by-snapshot.md index 54aa76008..444e9ba10 100644 --- a/en/restore-from-aws-s3-by-snapshot.md +++ b/en/restore-from-aws-s3-by-snapshot.md @@ -36,7 +36,7 @@ The restore method described in this document is implemented based on CustomReso Before using TiDB Operator to restore backup metadata and EBS snapshots from S3 storage to TiDB, prepare the restore environment by following the steps below: -1. Download the file [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml). +1. Download the file [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml). 2. Create the RBAC-related resources required for the restore in the `test2` namespace by running the following command: diff --git a/en/restore-from-aws-s3-using-br.md b/en/restore-from-aws-s3-using-br.md index dc85c4ed8..fd9c261a1 100644 --- a/en/restore-from-aws-s3-using-br.md +++ b/en/restore-from-aws-s3-using-br.md @@ -39,7 +39,7 @@ Before restoring backup data on a S3-compatible storage to TiDB using BR, take t kubectl create namespace restore-test ``` -2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: +2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: {{< copyable "shell-regular" >}} @@ -248,7 +248,7 @@ Before restoring backup data on S3-compatible storages to TiDB using BR, take th kubectl create namespace restore-test ``` -2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: +2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: ```shell kubectl apply -f backup-rbac.yaml -n restore-test diff --git a/en/restore-from-azblob-using-br.md b/en/restore-from-azblob-using-br.md index 5309b5074..26d80bfae 100644 --- a/en/restore-from-azblob-using-br.md +++ b/en/restore-from-azblob-using-br.md @@ -38,7 +38,7 @@ Before restoring backup data on Azure Blob Storage to TiDB using BR, take the fo kubectl create namespace restore-test ``` -2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: +2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: ```shell kubectl apply -f backup-rbac.yaml -n restore-test @@ -155,7 +155,7 @@ Before restoring backup data on Azure Blob Storage to TiDB using BR, take the fo kubectl create namespace restore-test ``` -2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: +2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: ```shell kubectl apply -f backup-rbac.yaml -n restore-test diff --git a/en/restore-from-gcs-using-br.md b/en/restore-from-gcs-using-br.md index 30da7b92b..0cb96f17e 100644 --- a/en/restore-from-gcs-using-br.md +++ b/en/restore-from-gcs-using-br.md @@ -40,7 +40,7 @@ Before restoring backup data on GCS to TiDB using BR, take the following steps t kubectl create namespace restore-test ``` -2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: +2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: ```shell kubectl apply -f backup-rbac.yaml -n restore-test @@ -160,7 +160,7 @@ Before restoring backup data on GCS to TiDB using BR, take the following steps t kubectl create namespace restore-test ``` -2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: +2. Download [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml), and execute the following command to create the role-based access control (RBAC) resources in the `restore-test` namespace: ```shell kubectl apply -f backup-rbac.yaml -n restore-test diff --git a/en/restore-from-gcs.md b/en/restore-from-gcs.md index 25cffac7f..6f25aed05 100644 --- a/en/restore-from-gcs.md +++ b/en/restore-from-gcs.md @@ -28,7 +28,7 @@ Before you perform the data restore, you need to prepare the restore environment ### Prepare the restore environment -1. Download [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml) and execute the following command to create the role-based access control (RBAC) resources in the `test2` namespace: +1. Download [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml) and execute the following command to create the role-based access control (RBAC) resources in the `test2` namespace: {{< copyable "shell-regular" >}} diff --git a/en/restore-from-pv-using-br.md b/en/restore-from-pv-using-br.md index b1e443475..ecf089868 100644 --- a/en/restore-from-pv-using-br.md +++ b/en/restore-from-pv-using-br.md @@ -22,7 +22,7 @@ After backing up TiDB cluster data to PVs using BR, if you need to recover the b Before restoring backup data on PVs to TiDB using BR, take the following steps to prepare the restore environment: -1. Download [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml). +1. Download [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml). 2. Execute the following command to create the role-based access control (RBAC) resources in the `test2` namespace: diff --git a/en/restore-from-s3.md b/en/restore-from-s3.md index aeeb5f583..d2a99ee59 100644 --- a/en/restore-from-s3.md +++ b/en/restore-from-s3.md @@ -28,7 +28,7 @@ Before you perform the data restore, you need to prepare the restore environment ### Prepare the restore environment -1. Download [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml) and execute the following command to create the role-based access control (RBAC) resources in the `test2` namespace: +1. Download [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml) and execute the following command to create the role-based access control (RBAC) resources in the `test2` namespace: {{< copyable "shell-regular" >}} diff --git a/en/tidb-toolkit.md b/en/tidb-toolkit.md index acc90bcdd..6602e058a 100644 --- a/en/tidb-toolkit.md +++ b/en/tidb-toolkit.md @@ -201,12 +201,12 @@ helm search repo pingcap ``` NAME CHART VERSION APP VERSION DESCRIPTION -pingcap/tidb-backup v1.5.0 A Helm chart for TiDB Backup or Restore -pingcap/tidb-cluster v1.5.0 A Helm chart for TiDB Cluster -pingcap/tidb-drainer v1.5.0 A Helm chart for TiDB Binlog drainer. -pingcap/tidb-lightning v1.5.0 A Helm chart for TiDB Lightning -pingcap/tidb-operator v1.5.0 v1.5.0 tidb-operator Helm chart for Kubernetes -pingcap/tikv-importer v1.5.0 A Helm chart for TiKV Importer +pingcap/tidb-backup v1.5.1 A Helm chart for TiDB Backup or Restore +pingcap/tidb-cluster v1.5.1 A Helm chart for TiDB Cluster +pingcap/tidb-drainer v1.5.1 A Helm chart for TiDB Binlog drainer. +pingcap/tidb-lightning v1.5.1 A Helm chart for TiDB Lightning +pingcap/tidb-operator v1.5.1 v1.5.1 tidb-operator Helm chart for Kubernetes +pingcap/tikv-importer v1.5.1 A Helm chart for TiKV Importer ``` When a new version of chart has been released, you can use `helm repo update` to update the repository cached locally: @@ -268,9 +268,9 @@ Use the following command to download the chart file required for cluster instal {{< copyable "shell-regular" >}} ```shell -wget http://charts.pingcap.org/tidb-operator-v1.5.0.tgz -wget http://charts.pingcap.org/tidb-drainer-v1.5.0.tgz -wget http://charts.pingcap.org/tidb-lightning-v1.5.0.tgz +wget http://charts.pingcap.org/tidb-operator-v1.5.1.tgz +wget http://charts.pingcap.org/tidb-drainer-v1.5.1.tgz +wget http://charts.pingcap.org/tidb-lightning-v1.5.1.tgz ``` Copy these chart files to the server and decompress them. You can use these charts to install the corresponding components by running the `helm install` command. Take `tidb-operator` as an example: @@ -278,7 +278,7 @@ Copy these chart files to the server and decompress them. You can use these char {{< copyable "shell-regular" >}} ```shell -tar zxvf tidb-operator.v1.5.0.tgz +tar zxvf tidb-operator.v1.5.1.tgz helm install ${release_name} ./tidb-operator --namespace=${namespace} ``` diff --git a/en/upgrade-tidb-operator.md b/en/upgrade-tidb-operator.md index 92566a276..7f021489d 100644 --- a/en/upgrade-tidb-operator.md +++ b/en/upgrade-tidb-operator.md @@ -60,27 +60,27 @@ If your server has access to the internet, you can perform online upgrade by tak kubectl get crd tidbclusters.pingcap.com ``` - This document takes TiDB v1.5.0 as an example. You can replace `${operator_version}` with the specific version you want to upgrade to. + This document takes TiDB v1.5.1 as an example. You can replace `${operator_version}` with the specific version you want to upgrade to. 3. Get the `values.yaml` file of the `tidb-operator` chart: {{< copyable "shell-regular" >}} ```bash - mkdir -p ${HOME}/tidb-operator/v1.5.0 && \ - helm inspect values pingcap/tidb-operator --version=v1.5.0 > ${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml + mkdir -p ${HOME}/tidb-operator/v1.5.1 && \ + helm inspect values pingcap/tidb-operator --version=v1.5.1 > ${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml ``` -4. In the `${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml` file, modify the `operatorImage` version to the new TiDB Operator version. +4. In the `${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml` file, modify the `operatorImage` version to the new TiDB Operator version. -5. If you have added customized configuration in the old `values.yaml` file, merge your customized configuration to the `${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml` file. +5. If you have added customized configuration in the old `values.yaml` file, merge your customized configuration to the `${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml` file. 6. Perform upgrade: {{< copyable "shell-regular" >}} ```bash - helm upgrade tidb-operator pingcap/tidb-operator --version=v1.5.0 -f ${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml -n tidb-admin + helm upgrade tidb-operator pingcap/tidb-operator --version=v1.5.1 -f ${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml -n tidb-admin ``` 7. After all the Pods start normally, check the image of TiDB Operator: @@ -91,13 +91,13 @@ If your server has access to the internet, you can perform online upgrade by tak kubectl get po -n tidb-admin -l app.kubernetes.io/instance=tidb-operator -o yaml | grep 'image:.*operator:' ``` - If you see a similar output as follows, TiDB Operator is successfully upgraded. `v1.5.0` represents the TiDB Operator version you have upgraded to. + If you see a similar output as follows, TiDB Operator is successfully upgraded. `v1.5.1` represents the TiDB Operator version you have upgraded to. ``` - image: pingcap/tidb-operator:v1.5.0 - image: docker.io/pingcap/tidb-operator:v1.5.0 - image: pingcap/tidb-operator:v1.5.0 - image: docker.io/pingcap/tidb-operator:v1.5.0 + image: pingcap/tidb-operator:v1.5.1 + image: docker.io/pingcap/tidb-operator:v1.5.1 + image: pingcap/tidb-operator:v1.5.1 + image: docker.io/pingcap/tidb-operator:v1.5.1 ``` ## Offline upgrade @@ -124,14 +124,14 @@ If your server cannot access the Internet, you can offline upgrade by taking the wget -O crd.yaml https://raw.githubusercontent.com/pingcap/tidb-operator/${operator_version}/manifests/crd_v1beta1.yaml ``` - This document takes TiDB v1.5.0 as an example. You can replace `${operator_version}` with the specific version you want to upgrade to. + This document takes TiDB v1.5.1 as an example. You can replace `${operator_version}` with the specific version you want to upgrade to. 2. Download the `tidb-operator` chart package file. {{< copyable "shell-regular" >}} ```bash - wget http://charts.pingcap.org/tidb-operator-v1.5.0.tgz + wget http://charts.pingcap.org/tidb-operator-v1.5.1.tgz ``` 3. Download the Docker images required for the new TiDB Operator version: @@ -139,11 +139,11 @@ If your server cannot access the Internet, you can offline upgrade by taking the {{< copyable "shell-regular" >}} ```bash - docker pull pingcap/tidb-operator:v1.5.0 - docker pull pingcap/tidb-backup-manager:v1.5.0 + docker pull pingcap/tidb-operator:v1.5.1 + docker pull pingcap/tidb-backup-manager:v1.5.1 - docker save -o tidb-operator-v1.5.0.tar pingcap/tidb-operator:v1.5.0 - docker save -o tidb-backup-manager-v1.5.0.tar pingcap/tidb-backup-manager:v1.5.0 + docker save -o tidb-operator-v1.5.1.tar pingcap/tidb-operator:v1.5.1 + docker save -o tidb-backup-manager-v1.5.1.tar pingcap/tidb-backup-manager:v1.5.1 ``` 2. Upload the downloaded files and images to the server where TiDB Operator is deployed, and install the new TiDB Operator version: @@ -171,9 +171,9 @@ If your server cannot access the Internet, you can offline upgrade by taking the {{< copyable "shell-regular" >}} ```bash - tar zxvf tidb-operator-v1.5.0.tgz && \ - mkdir -p ${HOME}/tidb-operator/v1.5.0 && \ - cp tidb-operator/values.yaml ${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml + tar zxvf tidb-operator-v1.5.1.tgz && \ + mkdir -p ${HOME}/tidb-operator/v1.5.1 && \ + cp tidb-operator/values.yaml ${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml ``` 4. Install the Docker images on the server: @@ -181,20 +181,20 @@ If your server cannot access the Internet, you can offline upgrade by taking the {{< copyable "shell-regular" >}} ```bash - docker load -i tidb-operator-v1.5.0.tar && \ - docker load -i tidb-backup-manager-v1.5.0.tar + docker load -i tidb-operator-v1.5.1.tar && \ + docker load -i tidb-backup-manager-v1.5.1.tar ``` -3. In the `${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml` file, modify the `operatorImage` version to the new TiDB Operator version. +3. In the `${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml` file, modify the `operatorImage` version to the new TiDB Operator version. -4. If you have added customized configuration in the old `values.yaml` file, merge your customized configuration to the `${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml` file. +4. If you have added customized configuration in the old `values.yaml` file, merge your customized configuration to the `${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml` file. 5. Perform upgrade: {{< copyable "shell-regular" >}} ```bash - helm upgrade tidb-operator ./tidb-operator --version=v1.5.0 -f ${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml + helm upgrade tidb-operator ./tidb-operator --version=v1.5.1 -f ${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml ``` 6. After all the Pods start normally, check the image version of TiDB Operator: @@ -205,13 +205,13 @@ If your server cannot access the Internet, you can offline upgrade by taking the kubectl get po -n tidb-admin -l app.kubernetes.io/instance=tidb-operator -o yaml | grep 'image:.*operator:' ``` - If you see a similar output as follows, TiDB Operator is successfully upgraded. `v1.5.0` represents the TiDB Operator version you have upgraded to. + If you see a similar output as follows, TiDB Operator is successfully upgraded. `v1.5.1` represents the TiDB Operator version you have upgraded to. ``` - image: pingcap/tidb-operator:v1.5.0 - image: docker.io/pingcap/tidb-operator:v1.5.0 - image: pingcap/tidb-operator:v1.5.0 - image: docker.io/pingcap/tidb-operator:v1.5.0 + image: pingcap/tidb-operator:v1.5.1 + image: docker.io/pingcap/tidb-operator:v1.5.1 + image: pingcap/tidb-operator:v1.5.1 + image: docker.io/pingcap/tidb-operator:v1.5.1 ``` > **Note:** diff --git a/zh/TOC.md b/zh/TOC.md index 43e82ead2..423a2c3b5 100644 --- a/zh/TOC.md +++ b/zh/TOC.md @@ -113,7 +113,7 @@ - [增强型 StatefulSet 控制器](advanced-statefulset.md) - [准入控制器](enable-admission-webhook.md) - [Sysbench 性能测试](benchmark-sysbench.md) - - [API 参考文档](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md) + - [API 参考文档](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md) - [Cheat Sheet](cheat-sheet.md) - [TiDB Operator RBAC 规则](tidb-operator-rbac.md) - 工具 diff --git a/zh/_index.md b/zh/_index.md index cba35262e..e0977985b 100644 --- a/zh/_index.md +++ b/zh/_index.md @@ -70,7 +70,7 @@ hide_commit: true -[API 参考文档](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md) +[API 参考文档](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md) [工具](https://docs.pingcap.com/zh/tidb-in-kubernetes/dev/tidb-toolkit) diff --git a/zh/access-dashboard.md b/zh/access-dashboard.md index c0a7f724f..97560a0c4 100644 --- a/zh/access-dashboard.md +++ b/zh/access-dashboard.md @@ -241,7 +241,7 @@ spec: EOF ``` - 关于 TidbNGMonitoring CR 的更多配置项,可参考 [tidb-operator 中的示例](https://github.com/pingcap/tidb-operator/blob/v1.5.0/examples/advanced/tidb-ng-monitoring.yaml)。 + 关于 TidbNGMonitoring CR 的更多配置项,可参考 [tidb-operator 中的示例](https://github.com/pingcap/tidb-operator/blob/v1.5.1/examples/advanced/tidb-ng-monitoring.yaml)。 3. 启用持续性能分析。 diff --git a/zh/advanced-statefulset.md b/zh/advanced-statefulset.md index ae9b130c0..068725697 100644 --- a/zh/advanced-statefulset.md +++ b/zh/advanced-statefulset.md @@ -21,7 +21,7 @@ Kubernetes 内置 [StatefulSet](https://kubernetes.io/docs/concepts/workloads/co {{< copyable "shell-regular" >}} ```shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/advanced-statefulset-crd.v1beta1.yaml + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/advanced-statefulset-crd.v1beta1.yaml ``` * Kubernetes 1.16 及之后版本: @@ -29,7 +29,7 @@ Kubernetes 内置 [StatefulSet](https://kubernetes.io/docs/concepts/workloads/co {{< copyable "shell-regular" >}} ```shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/advanced-statefulset-crd.v1.yaml + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/advanced-statefulset-crd.v1.yaml ``` 2. 在 TiDB Operator chart 的 `values.yaml` 中启用 `AdvancedStatefulSet` 特性: diff --git a/zh/aggregate-multiple-cluster-monitor-data.md b/zh/aggregate-multiple-cluster-monitor-data.md index 31598d80a..b31622266 100644 --- a/zh/aggregate-multiple-cluster-monitor-data.md +++ b/zh/aggregate-multiple-cluster-monitor-data.md @@ -24,7 +24,7 @@ Thanos 提供了跨 Prometheus 的统一查询方案 [Thanos Query](https://than {{< copyable "shell-regular" >}} ``` - kubectl -n ${namespace} apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/monitor-with-thanos/tidb-monitor.yaml + kubectl -n ${namespace} apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/monitor-with-thanos/tidb-monitor.yaml ``` 2. 部署 Thanos Query 组件。 @@ -34,7 +34,7 @@ Thanos 提供了跨 Prometheus 的统一查询方案 [Thanos Query](https://than {{< copyable "shell-regular" >}} ``` - curl -sl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/monitor-with-thanos/thanos-query.yaml + curl -sl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/monitor-with-thanos/thanos-query.yaml ``` 2. 手动修改 `thanos-query.yaml` 文件中的 `--store` 参数,将 `basic-prometheus:10901` 改为 `basic-prometheus.${namespace}:10901`。 @@ -182,4 +182,4 @@ spec: Prometheus 将会把数据推送到 [Thanos Receiver](https://thanos.io/tip/components/receive.md/) 服务,详情可以参考 [Receiver 架构设计](https://thanos.io/v0.8/proposals/201812_thanos-remote-receive/)。 -部署方案可以参考 [Example](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-prom-remotewrite)。 +部署方案可以参考 [Example](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-prom-remotewrite)。 diff --git a/zh/backup-restore-cr.md b/zh/backup-restore-cr.md index a2aeb1cef..84236db58 100644 --- a/zh/backup-restore-cr.md +++ b/zh/backup-restore-cr.md @@ -22,7 +22,7 @@ summary: 介绍用于备份与恢复的 Custom Resource (CR) 资源的各字段 - 如果指定了镜像但未指定版本,例如 `.spec.toolImage: private/registry/br`,那么使用镜像 `private/registry/br:${tikv_version}` 进行备份。 - 使用 Dumpling 备份时,可以用该字段指定 Dumpling 的版本: - 如果指定了 Dumpling 的版本,例如 `spec.toolImage: pingcap/dumpling:v5.3.0`,那么使用指定的版本镜像进行备份。 - - 如果未指定,默认使用 [Backup Manager Dockerfile](https://github.com/pingcap/tidb-operator/blob/v1.5.0/images/tidb-backup-manager/Dockerfile) 文件中 `TOOLKIT_VERSION` 指定的 Dumpling 版本进行备份。 + - 如果未指定,默认使用 [Backup Manager Dockerfile](https://github.com/pingcap/tidb-operator/blob/v1.5.1/images/tidb-backup-manager/Dockerfile) 文件中 `TOOLKIT_VERSION` 指定的 Dumpling 版本进行备份。 * `.spec.backupType`:指定 Backup 类型,该字段仅在使用 BR 备份时有效,目前支持以下三种类型,可以结合 `.spec.tableFilter` 配置表库过滤规则: * `full`:对 TiDB 集群所有的 database 数据执行备份。 @@ -247,7 +247,7 @@ summary: 介绍用于备份与恢复的 Custom Resource (CR) 资源的各字段 * `.spec.metadata.namespace`:`Restore` CR 所在的 namespace。 * `.spec.toolImage`:用于指定 `Restore` 使用的工具镜像。TiDB Operator 从 v1.1.9 版本起支持这项配置。 - 使用 BR 恢复时,可以用该字段指定 BR 的版本。例如,`spec.toolImage: pingcap/br:v5.3.0`。如果不指定,默认使用 `pingcap/br:${tikv_version}` 进行恢复。 - - 使用 Lightning 恢复时,可以用该字段指定 Lightning 的版本,例如`spec.toolImage: pingcap/lightning:v5.3.0`。如果不指定,默认使用 [Backup Manager Dockerfile](https://github.com/pingcap/tidb-operator/blob/v1.5.0/images/tidb-backup-manager/Dockerfile) 文件中 `TOOLKIT_VERSION` 指定的 Lightning 版本进行恢复。 + - 使用 Lightning 恢复时,可以用该字段指定 Lightning 的版本,例如`spec.toolImage: pingcap/lightning:v5.3.0`。如果不指定,默认使用 [Backup Manager Dockerfile](https://github.com/pingcap/tidb-operator/blob/v1.5.1/images/tidb-backup-manager/Dockerfile) 文件中 `TOOLKIT_VERSION` 指定的 Lightning 版本进行恢复。 * `.spec.backupType`:指定 Restore 类型,该字段仅在使用 BR 恢复时有效,目前支持以下三种类型,可以结合 `.spec.tableFilter` 配置表库过滤规则: * `full`:对 TiDB 集群所有的 database 数据执行备份。 diff --git a/zh/backup-to-aws-s3-by-snapshot.md b/zh/backup-to-aws-s3-by-snapshot.md index 98e0f72a5..ae871f96f 100644 --- a/zh/backup-to-aws-s3-by-snapshot.md +++ b/zh/backup-to-aws-s3-by-snapshot.md @@ -42,7 +42,7 @@ summary: 介绍如何基于 EBS 卷快照使用 TiDB Operator 备份 TiDB 集群 ### 第 1 步:准备 EBS 卷快照备份环境 -1. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml) 到执行备份的服务器。 +1. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml) 到执行备份的服务器。 2. 执行以下命令,在 `test1` 这个命名空间中,创建备份需要的 RBAC 相关资源: diff --git a/zh/backup-to-aws-s3-using-br.md b/zh/backup-to-aws-s3-using-br.md index 475ceab8c..5163ac573 100644 --- a/zh/backup-to-aws-s3-using-br.md +++ b/zh/backup-to-aws-s3-using-br.md @@ -51,7 +51,7 @@ Ad-hoc 备份支持快照备份,也支持[启动](#启动日志备份)和[停 kubectl create namespace backup-test ``` -2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `backup-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: +2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `backup-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: {{< copyable "shell-regular" >}} diff --git a/zh/backup-to-azblob-using-br.md b/zh/backup-to-azblob-using-br.md index 016408a77..06b659426 100644 --- a/zh/backup-to-azblob-using-br.md +++ b/zh/backup-to-azblob-using-br.md @@ -48,7 +48,7 @@ Ad-hoc 备份支持快照备份,也支持[启动](#启动日志备份)和[停 kubectl create namespace backup-test ``` -2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `backup-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: +2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `backup-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: ```shell kubectl apply -f backup-rbac.yaml -n backup-test diff --git a/zh/backup-to-gcs-using-br.md b/zh/backup-to-gcs-using-br.md index 9d6260070..04b2eeb98 100644 --- a/zh/backup-to-gcs-using-br.md +++ b/zh/backup-to-gcs-using-br.md @@ -49,7 +49,7 @@ Ad-hoc 备份支持快照备份,也支持[启动](#启动日志备份)和[停 kubectl create namespace backup-test ``` -2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `backup-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: +2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `backup-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: {{< copyable "shell-regular" >}} diff --git a/zh/backup-to-gcs.md b/zh/backup-to-gcs.md index ec68e66d8..3e64080d3 100644 --- a/zh/backup-to-gcs.md +++ b/zh/backup-to-gcs.md @@ -38,7 +38,7 @@ Ad-hoc 全量备份通过创建一个自定义的 `Backup` custom resource (CR) ### 第 1 步:Ad-hoc 全量备份环境准备 -1. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `test1` 这个 namespace 中创建备份需要的 RBAC 相关资源: +1. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `test1` 这个 namespace 中创建备份需要的 RBAC 相关资源: {{< copyable "shell-regular" >}} diff --git a/zh/backup-to-pv-using-br.md b/zh/backup-to-pv-using-br.md index eaef57bef..3f5cf1d11 100644 --- a/zh/backup-to-pv-using-br.md +++ b/zh/backup-to-pv-using-br.md @@ -31,7 +31,7 @@ Ad-hoc 备份支持快照备份与增量备份。Ad-hoc 备份通过创建一个 ### 第 1 步:准备 Ad-hoc 备份环境 -1. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml) 到执行备份的服务器。 +1. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml) 到执行备份的服务器。 2. 执行以下命令,在 `test1` 这个命名空间中,创建备份需要的 RBAC 相关资源: diff --git a/zh/backup-to-s3.md b/zh/backup-to-s3.md index 0407ae42a..da220d7df 100644 --- a/zh/backup-to-s3.md +++ b/zh/backup-to-s3.md @@ -50,12 +50,12 @@ GRANT ### 第 1 步:Ad-hoc 全量备份环境准备 -1. 执行以下命令,根据 [backup-rbac.yaml](https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/backup/backup-rbac.yaml) 在 `tidb-cluster` 命名空间创建基于角色的访问控制 (RBAC) 资源。 +1. 执行以下命令,根据 [backup-rbac.yaml](https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/backup/backup-rbac.yaml) 在 `tidb-cluster` 命名空间创建基于角色的访问控制 (RBAC) 资源。 {{< copyable "shell-regular" >}} ```shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/backup/backup-rbac.yaml -n tidb-cluster + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/backup/backup-rbac.yaml -n tidb-cluster ``` 2. 远程存储访问授权。 diff --git a/zh/cheat-sheet.md b/zh/cheat-sheet.md index 363622b87..a62b1a3d6 100644 --- a/zh/cheat-sheet.md +++ b/zh/cheat-sheet.md @@ -493,7 +493,7 @@ helm inspect values ${chart_name} --version=${chart_version} > values.yaml {{< copyable "shell-regular" >}} ```shell -helm inspect values pingcap/tidb-operator --version=v1.5.0 > values-tidb-operator.yaml +helm inspect values pingcap/tidb-operator --version=v1.5.1 > values-tidb-operator.yaml ``` ### 使用 Helm Chart 部署 @@ -509,7 +509,7 @@ helm install ${name} ${chart_name} --namespace=${namespace} --version=${chart_ve {{< copyable "shell-regular" >}} ```shell -helm install tidb-operator pingcap/tidb-operator --namespace=tidb-admin --version=v1.5.0 -f values-tidb-operator.yaml +helm install tidb-operator pingcap/tidb-operator --namespace=tidb-admin --version=v1.5.1 -f values-tidb-operator.yaml ``` ### 查看已经部署的 Helm Release @@ -533,7 +533,7 @@ helm upgrade ${name} ${chart_name} --version=${chart_version} -f ${values_file} {{< copyable "shell-regular" >}} ```shell -helm upgrade tidb-operator pingcap/tidb-operator --version=v1.5.0 -f values-tidb-operator.yaml +helm upgrade tidb-operator pingcap/tidb-operator --version=v1.5.1 -f values-tidb-operator.yaml ``` ### 删除 Helm Release diff --git a/zh/configure-a-tidb-cluster.md b/zh/configure-a-tidb-cluster.md index 33e2758e2..d7223642b 100644 --- a/zh/configure-a-tidb-cluster.md +++ b/zh/configure-a-tidb-cluster.md @@ -25,7 +25,7 @@ aliases: ['/docs-cn/tidb-in-kubernetes/dev/configure-a-tidb-cluster/','/zh/tidb- ## 部署配置 -通过配置 `TidbCluster` CR 来配置 TiDB 集群。参考 TidbCluster [示例](https://github.com/pingcap/tidb-operator/blob/v1.5.0/examples/advanced/tidb-cluster.yaml)和 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md)(示例和 API 文档请切换到当前使用的 TiDB Operator 版本)完成 TidbCluster CR(Custom Resource)。 +通过配置 `TidbCluster` CR 来配置 TiDB 集群。参考 TidbCluster [示例](https://github.com/pingcap/tidb-operator/blob/v1.5.1/examples/advanced/tidb-cluster.yaml)和 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md)(示例和 API 文档请切换到当前使用的 TiDB Operator 版本)完成 TidbCluster CR(Custom Resource)。 > **注意:** > diff --git a/zh/configure-storage-class.md b/zh/configure-storage-class.md index 84cef1a23..a020a9624 100644 --- a/zh/configure-storage-class.md +++ b/zh/configure-storage-class.md @@ -103,7 +103,7 @@ Kubernetes 当前支持静态分配的本地存储。可使用 [local-static-pro {{< copyable "shell-regular" >}} ```shell - wget https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/local-pv/local-volume-provisioner.yaml + wget https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/local-pv/local-volume-provisioner.yaml ``` 2. 如果你使用的发现路径与[第 1 步:准备本地存储](#第-1-步准备本地存储)中的示例一致,可跳过这一步。如果你使用与上一步中不同路径的发现目录,需要修改 ConfigMap 和 DaemonSet 定义。 diff --git a/zh/deploy-heterogeneous-tidb-cluster.md b/zh/deploy-heterogeneous-tidb-cluster.md index 4ee74be9c..837388075 100644 --- a/zh/deploy-heterogeneous-tidb-cluster.md +++ b/zh/deploy-heterogeneous-tidb-cluster.md @@ -165,7 +165,7 @@ summary: 本文档介绍如何为已有的 TiDB 集群部署一个异构集群 其中,`spec.tlsCluster.enabled` 表示组件间是否开启 TLS,`spec.tidb.tlsClient.enabled` 表示 MySQL 客户端是否开启 TLS。 - - 详细的异构 TLS 集群配置示例,请参阅 [`heterogeneous-tls`](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/heterogeneous-tls)。 + - 详细的异构 TLS 集群配置示例,请参阅 [`heterogeneous-tls`](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/heterogeneous-tls)。 - TiDB 集群更多的配置项和字段含义,请参考 [TiDB 集群配置文档](configure-a-tidb-cluster.md)。 diff --git a/zh/deploy-on-alibaba-cloud.md b/zh/deploy-on-alibaba-cloud.md index e33b48049..bbfd113fd 100644 --- a/zh/deploy-on-alibaba-cloud.md +++ b/zh/deploy-on-alibaba-cloud.md @@ -89,7 +89,7 @@ aliases: ['/docs-cn/tidb-in-kubernetes/dev/deploy-on-alibaba-cloud/'] tikv_count = 3 tidb_count = 2 pd_count = 3 - operator_version = "v1.5.0" + operator_version = "v1.5.1" ``` 如果需要在集群中部署 TiFlash,需要在 `terraform.tfvars` 中设置 `create_tiflash_node_pool = true`,也可以设置 `tiflash_count` 和 `tiflash_instance_type` 来配置 TiFlash 节点池的节点数量和实例类型,`tiflash_count` 默认为 `2`,`tiflash_instance_type` 默认为 `ecs.i2.2xlarge`。 @@ -168,7 +168,7 @@ aliases: ['/docs-cn/tidb-in-kubernetes/dev/deploy-on-alibaba-cloud/'] cp manifests/dashboard.yaml.example tidb-dashboard.yaml ``` - 参考 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md)和[集群配置文档](configure-a-tidb-cluster.md)完成 CR 文件配置。 + 参考 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md)和[集群配置文档](configure-a-tidb-cluster.md)完成 CR 文件配置。 如果要部署 TiFlash,可以在 db.yaml 中配置 `spec.tiflash`,例如: @@ -374,7 +374,7 @@ terraform state rm module.ack.alicloud_cs_managed_kubernetes.k8s ### 配置 TiDB 集群 -参考 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md)和[集群配置文档](configure-a-tidb-cluster.md)修改 TiDB 集群配置。 +参考 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md)和[集群配置文档](configure-a-tidb-cluster.md)修改 TiDB 集群配置。 ## 管理多个 TiDB 集群 diff --git a/zh/deploy-on-aws-eks.md b/zh/deploy-on-aws-eks.md index bd1138b66..1359fd110 100644 --- a/zh/deploy-on-aws-eks.md +++ b/zh/deploy-on-aws-eks.md @@ -296,7 +296,7 @@ mountOptions: 2. 通过[普通挂载方式](https://github.com/kubernetes-sigs/sig-storage-local-static-provisioner/blob/master/docs/operations.md#use-a-whole-disk-as-a-filesystem-pv)将本地存储挂载到 `/mnt/ssd` 目录。 - 3. 根据本地存储的挂载情况,修改 [local-volume-provisioner.yaml](https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/eks/local-volume-provisioner.yaml) 文件。 + 3. 根据本地存储的挂载情况,修改 [local-volume-provisioner.yaml](https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/eks/local-volume-provisioner.yaml) 文件。 4. 使用修改后的 `local-volume-provisioner.yaml`,部署并创建一个 `local-storage` 的 Storage Class: @@ -341,9 +341,9 @@ kubectl create namespace tidb-cluster {{< copyable "shell-regular" >}} ```shell -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aws/tidb-cluster.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aws/tidb-monitor.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aws/tidb-dashboard.yaml +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aws/tidb-cluster.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aws/tidb-monitor.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aws/tidb-dashboard.yaml ``` 如需了解更详细的配置信息或者进行自定义配置,请参考[配置 TiDB 集群](configure-a-tidb-cluster.md) @@ -650,4 +650,4 @@ spec: 最后使用 `kubectl -n tidb-cluster apply -f tidb-cluster.yaml` 更新 TiDB 集群配置。 -更多可参考 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md)和[集群配置文档](configure-a-tidb-cluster.md)完成 CR 文件配置。 +更多可参考 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md)和[集群配置文档](configure-a-tidb-cluster.md)完成 CR 文件配置。 diff --git a/zh/deploy-on-azure-aks.md b/zh/deploy-on-azure-aks.md index 0dec78bea..1c7af6cd1 100644 --- a/zh/deploy-on-azure-aks.md +++ b/zh/deploy-on-azure-aks.md @@ -233,9 +233,9 @@ kubectl create namespace tidb-cluster {{< copyable "shell-regular" >}} ```shell -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aks/tidb-cluster.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aks/tidb-monitor.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/aks/tidb-dashboard.yaml +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aks/tidb-cluster.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aks/tidb-monitor.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/aks/tidb-dashboard.yaml ``` 如需了解更详细的配置信息或者进行自定义配置,请参考[配置 TiDB 集群](configure-a-tidb-cluster.md) @@ -514,7 +514,7 @@ spec: 最后使用 `kubectl -n tidb-cluster apply -f tidb-cluster.yaml` 更新 TiDB 集群配置。 -更多可参考 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md)和[集群配置文档](configure-a-tidb-cluster.md)完成 CR 文件配置。 +更多可参考 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md)和[集群配置文档](configure-a-tidb-cluster.md)完成 CR 文件配置。 ## 使用其他 Azure 磁盘类型 @@ -593,7 +593,7 @@ Azure Disk 支持多种磁盘类型。若需要低延迟、高吞吐,可以选 {{< copyable "shell-regular" >}} ```shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/eks/local-volume-provisioner.yaml + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/eks/local-volume-provisioner.yaml ``` 3. 使用本地存储。 diff --git a/zh/deploy-on-gcp-gke.md b/zh/deploy-on-gcp-gke.md index ddbadd18c..4017e87b4 100644 --- a/zh/deploy-on-gcp-gke.md +++ b/zh/deploy-on-gcp-gke.md @@ -130,7 +130,7 @@ mountOptions: {{< copyable "shell-regular" >}} ```shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/gke/local-ssd-provision/local-ssd-provision.yaml + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/gke/local-ssd-provision/local-ssd-provision.yaml ``` 3. 使用本地存储。 @@ -166,9 +166,9 @@ kubectl create namespace tidb-cluster {{< copyable "shell-regular" >}} ```shell -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/gcp/tidb-cluster.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/gcp/tidb-monitor.yaml && \ -curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/gcp/tidb-dashboard.yaml +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/gcp/tidb-cluster.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/gcp/tidb-monitor.yaml && \ +curl -O https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/gcp/tidb-dashboard.yaml ``` 如需了解更详细的配置信息或者进行自定义配置,请参考[配置 TiDB 集群](configure-a-tidb-cluster.md) @@ -446,4 +446,4 @@ spec: 最后使用 `kubectl -n tidb-cluster apply -f tidb-cluster.yaml` 更新 TiDB 集群配置。 -更多可参考 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md)和[集群配置文档](configure-a-tidb-cluster.md)完成 CR 文件配置。 +更多可参考 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md)和[集群配置文档](configure-a-tidb-cluster.md)完成 CR 文件配置。 diff --git a/zh/deploy-tidb-cluster-across-multiple-kubernetes.md b/zh/deploy-tidb-cluster-across-multiple-kubernetes.md index 5c16b2de3..e9264fa16 100644 --- a/zh/deploy-tidb-cluster-across-multiple-kubernetes.md +++ b/zh/deploy-tidb-cluster-across-multiple-kubernetes.md @@ -639,7 +639,7 @@ kubectl patch tidbcluster cluster1 --type merge -p '{"spec":{"acrossK8s": true}} 完成上述步骤后,该 TidbCluster 可以作为跨 Kubernetes 集群部署 TiDB 集群的初始 TidbCluster。可以参考[部署新的 TidbCluster 加入 TiDB 集群](#第-2-步部署新的-tidbcluster-加入-tidb-集群)一节部署其他的 TidbCluster。 -更多示例信息以及开发信息,请参阅 [`multi-cluster`](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/multi-cluster)。 +更多示例信息以及开发信息,请参阅 [`multi-cluster`](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/multi-cluster)。 ## 跨多个 Kubernetes 集群部署的 TiDB 集群监控 diff --git a/zh/deploy-tidb-dm.md b/zh/deploy-tidb-dm.md index 329b36da2..cada744c8 100644 --- a/zh/deploy-tidb-dm.md +++ b/zh/deploy-tidb-dm.md @@ -17,7 +17,7 @@ summary: 了解如何在 Kubernetes 上部署 TiDB DM 集群。 ## 部署配置 -通过配置 DMCluster CR 来配置 DM 集群。参考 DMCluster [示例](https://github.com/pingcap/tidb-operator/blob/v1.5.0/examples/dm/dm-cluster.yaml)和 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md#dmcluster)(示例和 API 文档请切换到当前使用的 TiDB Operator 版本)完成 DMCluster CR (Custom Resource)。 +通过配置 DMCluster CR 来配置 DM 集群。参考 DMCluster [示例](https://github.com/pingcap/tidb-operator/blob/v1.5.1/examples/dm/dm-cluster.yaml)和 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md#dmcluster)(示例和 API 文档请切换到当前使用的 TiDB Operator 版本)完成 DMCluster CR (Custom Resource)。 ### 集群名称 diff --git a/zh/deploy-tidb-from-kubernetes-gke.md b/zh/deploy-tidb-from-kubernetes-gke.md index b7179d915..b5955a699 100644 --- a/zh/deploy-tidb-from-kubernetes-gke.md +++ b/zh/deploy-tidb-from-kubernetes-gke.md @@ -94,7 +94,7 @@ kubectl get nodes TiDB Operator 使用 [Custom Resource Definition (CRD)](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#customresourcedefinitions) 扩展 Kubernetes,所以要使用 TiDB Operator,必须先创建 `TidbCluster` 等各种自定义资源类型: ```shell -kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/crd.yaml && \ +kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/crd.yaml && \ kubectl get crd tidbclusters.pingcap.com ``` @@ -106,7 +106,7 @@ kubectl get crd tidbclusters.pingcap.com ```shell kubectl create namespace tidb-admin -helm install --namespace tidb-admin tidb-operator pingcap/tidb-operator --version v1.5.0 +helm install --namespace tidb-admin tidb-operator pingcap/tidb-operator --version v1.5.1 kubectl get po -n tidb-admin -l app.kubernetes.io/name=tidb-operator ``` @@ -123,13 +123,13 @@ kubectl get po -n tidb-admin -l app.kubernetes.io/name=tidb-operator 2. 部署 TiDB 集群: ``` shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic/tidb-cluster.yaml -n demo + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic/tidb-cluster.yaml -n demo ``` 3. 部署 TiDB 集群监控: ``` shell - kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic/tidb-monitor.yaml -n demo + kubectl apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic/tidb-monitor.yaml -n demo ``` 4. 通过下面命令查看 Pod 状态: diff --git a/zh/deploy-tidb-monitor-across-multiple-kubernetes.md b/zh/deploy-tidb-monitor-across-multiple-kubernetes.md index 0e1cd16bd..e582f91ce 100644 --- a/zh/deploy-tidb-monitor-across-multiple-kubernetes.md +++ b/zh/deploy-tidb-monitor-across-multiple-kubernetes.md @@ -24,7 +24,7 @@ Push 方式指利用 Prometheus remote-write 的特性,使位于不同 Kuberne - 各 Kubernetes 集群上的 Prometheus(即 TidbMonitor)组件有能力访问 Thanos Receiver 组件。 -关于 Thanos Receiver 部署,可参考 [kube-thanos](https://github.com/thanos-io/kube-thanos) 以及 [Example](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-prom-remotewrite)。 +关于 Thanos Receiver 部署,可参考 [kube-thanos](https://github.com/thanos-io/kube-thanos) 以及 [Example](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-prom-remotewrite)。 ### 部署架构图 @@ -108,7 +108,7 @@ Pull 方式是指从不同 Kubernetes 集群的 Prometheus 实例中拉取监控 - Thanos Query 组件有能力访问各 Kubernetes 集群上的 Prometheus (即 TidbMonitor) 组件的 Pod IP。 - Thanos Query 组件有能力访问各 Kubernetes 集群上的 Prometheus (即 TidbMonitor) 组件的 Pod FQDN。 -关于 Thanos Query 部署, 参考 [kube-thanos](https://github.com/thanos-io/kube-thanos) 以及 [Example](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-with-thanos)。 +关于 Thanos Query 部署, 参考 [kube-thanos](https://github.com/thanos-io/kube-thanos) 以及 [Example](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-with-thanos)。 #### 部署架构图 diff --git a/zh/deploy-tidb-operator.md b/zh/deploy-tidb-operator.md index 87c56d9e7..fa08652ee 100644 --- a/zh/deploy-tidb-operator.md +++ b/zh/deploy-tidb-operator.md @@ -45,7 +45,7 @@ TiDB Operator 使用 [Custom Resource Definition (CRD)](https://kubernetes.io/do {{< copyable "shell-regular" >}} ```shell -kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/crd.yaml +kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/crd.yaml ``` 如果服务器没有外网,需要先用有外网的机器下载 `crd.yaml` 文件,然后再进行安装: @@ -53,7 +53,7 @@ kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0 {{< copyable "shell-regular" >}} ```shell -wget https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/crd.yaml +wget https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/crd.yaml kubectl create -f ./crd.yaml ``` @@ -101,7 +101,7 @@ tidbmonitors.pingcap.com 2020-06-11T07:59:41Z > **注意:** > - > `${chart_version}` 在后续文档中代表 chart 版本,例如 `v1.5.0`,可以通过 `helm search repo -l tidb-operator` 查看当前支持的版本。 + > `${chart_version}` 在后续文档中代表 chart 版本,例如 `v1.5.1`,可以通过 `helm search repo -l tidb-operator` 查看当前支持的版本。 2. 配置 TiDB Operator @@ -151,15 +151,15 @@ tidbmonitors.pingcap.com 2020-06-11T07:59:41Z {{< copyable "shell-regular" >}} ```shell - wget http://charts.pingcap.org/tidb-operator-v1.5.0.tgz + wget http://charts.pingcap.org/tidb-operator-v1.5.1.tgz ``` - 将 `tidb-operator-v1.5.0.tgz` 文件拷贝到服务器上并解压到当前目录: + 将 `tidb-operator-v1.5.1.tgz` 文件拷贝到服务器上并解压到当前目录: {{< copyable "shell-regular" >}} ```shell - tar zxvf tidb-operator.v1.5.0.tgz + tar zxvf tidb-operator.v1.5.1.tgz ``` 2. 下载 TiDB Operator 运行所需的 Docker 镜像 @@ -169,8 +169,8 @@ tidbmonitors.pingcap.com 2020-06-11T07:59:41Z TiDB Operator 用到的 Docker 镜像有: ```shell - pingcap/tidb-operator:v1.5.0 - pingcap/tidb-backup-manager:v1.5.0 + pingcap/tidb-operator:v1.5.1 + pingcap/tidb-backup-manager:v1.5.1 bitnami/kubectl:latest pingcap/advanced-statefulset:v0.3.3 k8s.gcr.io/kube-scheduler:v1.16.9 @@ -183,13 +183,13 @@ tidbmonitors.pingcap.com 2020-06-11T07:59:41Z {{< copyable "shell-regular" >}} ```shell - docker pull pingcap/tidb-operator:v1.5.0 - docker pull pingcap/tidb-backup-manager:v1.5.0 + docker pull pingcap/tidb-operator:v1.5.1 + docker pull pingcap/tidb-backup-manager:v1.5.1 docker pull bitnami/kubectl:latest docker pull pingcap/advanced-statefulset:v0.3.3 - docker save -o tidb-operator-v1.5.0.tar pingcap/tidb-operator:v1.5.0 - docker save -o tidb-backup-manager-v1.5.0.tar pingcap/tidb-backup-manager:v1.5.0 + docker save -o tidb-operator-v1.5.1.tar pingcap/tidb-operator:v1.5.1 + docker save -o tidb-backup-manager-v1.5.1.tar pingcap/tidb-backup-manager:v1.5.1 docker save -o bitnami-kubectl.tar bitnami/kubectl:latest docker save -o advanced-statefulset-v0.3.3.tar pingcap/advanced-statefulset:v0.3.3 ``` @@ -199,8 +199,8 @@ tidbmonitors.pingcap.com 2020-06-11T07:59:41Z {{< copyable "shell-regular" >}} ```shell - docker load -i tidb-operator-v1.5.0.tar - docker load -i tidb-backup-manager-v1.5.0.tar + docker load -i tidb-operator-v1.5.1.tar + docker load -i tidb-backup-manager-v1.5.1.tar docker load -i bitnami-kubectl.tar docker load -i advanced-statefulset-v0.3.3.tar ``` diff --git a/zh/enable-monitor-dynamic-configuration.md b/zh/enable-monitor-dynamic-configuration.md index 71454dfa9..4807adcaa 100644 --- a/zh/enable-monitor-dynamic-configuration.md +++ b/zh/enable-monitor-dynamic-configuration.md @@ -39,7 +39,7 @@ spec: `prometheusReloader` 配置变更后,TidbMonitor 会自动重启。重启后,所有针对 Prometheus 的配置变更都会动态更新。 -可以参考 [monitor-dynamic-configmap 配置示例](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-dynamic-configmap)。 +可以参考 [monitor-dynamic-configmap 配置示例](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-dynamic-configmap)。 ## 关闭动态配置功能 diff --git a/zh/enable-monitor-shards.md b/zh/enable-monitor-shards.md index 792bd87f3..4f989e401 100644 --- a/zh/enable-monitor-shards.md +++ b/zh/enable-monitor-shards.md @@ -49,4 +49,4 @@ spec: > - TidbMonitor 对应的 Pod 实例数量取决于 `replicas` 和 `shards` 的乘积。例如,当 `replicas` 为 1 个副本,`shards` 为 2 个分片时,TiDB Operator 将产生 2 个 TidbMonitor Pod 实例。 > - `shards` 变更后,`Targets` 会重新分配,但是原本在节点上的监控数据不会重新分配。 -可以参考 [分片示例](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-shards)。 +可以参考 [分片示例](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-shards)。 diff --git a/zh/get-started.md b/zh/get-started.md index 0d65493a7..c6d03be5c 100644 --- a/zh/get-started.md +++ b/zh/get-started.md @@ -196,7 +196,7 @@ TiDB Operator 包含许多实现 TiDB 集群不同组件的自定义资源类型 {{< copyable "shell-regular" >}} ```shell -kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/manifests/crd.yaml +kubectl create -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/manifests/crd.yaml ```
@@ -261,7 +261,7 @@ customresourcedefinition.apiextensions.k8s.io/tidbclusterautoscalers.pingcap.com {{< copyable "shell-regular" >}} ```shell - helm install --namespace tidb-admin tidb-operator pingcap/tidb-operator --version v1.5.0 + helm install --namespace tidb-admin tidb-operator pingcap/tidb-operator --version v1.5.1 ``` 如果访问 Docker Hub 网速较慢,可以使用阿里云上的镜像: @@ -269,9 +269,9 @@ customresourcedefinition.apiextensions.k8s.io/tidbclusterautoscalers.pingcap.com {{< copyable "shell-regular" >}} ``` - helm install --namespace tidb-admin tidb-operator pingcap/tidb-operator --version v1.5.0 \ - --set operatorImage=registry.cn-beijing.aliyuncs.com/tidb/tidb-operator:v1.5.0 \ - --set tidbBackupManagerImage=registry.cn-beijing.aliyuncs.com/tidb/tidb-backup-manager:v1.5.0 \ + helm install --namespace tidb-admin tidb-operator pingcap/tidb-operator --version v1.5.1 \ + --set operatorImage=registry.cn-beijing.aliyuncs.com/tidb/tidb-operator:v1.5.1 \ + --set tidbBackupManagerImage=registry.cn-beijing.aliyuncs.com/tidb/tidb-backup-manager:v1.5.1 \ --set scheduler.kubeSchedulerImageName=registry.cn-hangzhou.aliyuncs.com/google_containers/kube-scheduler ``` @@ -324,7 +324,7 @@ tidb-scheduler-644d59b46f-4f6sb 2/2 Running 0 2m22s ``` shell kubectl create namespace tidb-cluster && \ - kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic/tidb-cluster.yaml + kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic/tidb-cluster.yaml ``` 如果访问 Docker Hub 网速较慢,可以使用 UCloud 上的镜像: @@ -333,7 +333,7 @@ kubectl create namespace tidb-cluster && \ ``` kubectl create namespace tidb-cluster && \ - kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic-cn/tidb-cluster.yaml + kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic-cn/tidb-cluster.yaml ```
@@ -353,7 +353,7 @@ tidbcluster.pingcap.com/basic created {{< copyable "shell-regular" >}} ``` shell -kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic/tidb-dashboard.yaml +kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic/tidb-dashboard.yaml ``` 如果访问 Docker Hub 网速较慢,可以使用 UCloud 上的镜像: @@ -361,7 +361,7 @@ kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb- {{< copyable "shell-regular" >}} ``` -kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic-cn/tidb-dashboard.yaml +kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic-cn/tidb-dashboard.yaml ```
@@ -378,7 +378,7 @@ tidbdashboard.pingcap.com/basic created {{< copyable "shell-regular" >}} ``` shell -kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic/tidb-monitor.yaml +kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic/tidb-monitor.yaml ``` 如果访问 Docker Hub 网速较慢,可以使用 UCloud 上的镜像: @@ -386,7 +386,7 @@ kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb- {{< copyable "shell-regular" >}} ``` -kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.0/examples/basic-cn/tidb-monitor.yaml +kubectl -n tidb-cluster apply -f https://raw.githubusercontent.com/pingcap/tidb-operator/v1.5.1/examples/basic-cn/tidb-monitor.yaml ```
diff --git a/zh/initialize-a-cluster.md b/zh/initialize-a-cluster.md index 49a73a45f..f3ed5f0bf 100644 --- a/zh/initialize-a-cluster.md +++ b/zh/initialize-a-cluster.md @@ -15,7 +15,7 @@ aliases: ['/docs-cn/tidb-in-kubernetes/dev/initialize-a-cluster/'] ## 配置 TidbInitializer -请参考 TidbInitializer [示例](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/initializer/tidb-initializer.yaml)和 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.0/docs/api-references/docs.md)(示例和 API 文档请切换到当前使用的 TiDB Operator 版本)以及下面的步骤,完成 TidbInitializer CR,保存到文件 `${cluster_name}/tidb-initializer.yaml`。 +请参考 TidbInitializer [示例](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/initializer/tidb-initializer.yaml)和 [API 文档](https://github.com/pingcap/tidb-operator/blob/v1.5.1/docs/api-references/docs.md)(示例和 API 文档请切换到当前使用的 TiDB Operator 版本)以及下面的步骤,完成 TidbInitializer CR,保存到文件 `${cluster_name}/tidb-initializer.yaml`。 ### 设置集群的命名空间和名称 diff --git a/zh/monitor-a-tidb-cluster.md b/zh/monitor-a-tidb-cluster.md index 8445332a0..5569c0b7a 100644 --- a/zh/monitor-a-tidb-cluster.md +++ b/zh/monitor-a-tidb-cluster.md @@ -14,7 +14,7 @@ TiDB 通过 Prometheus 和 Grafana 监控 TiDB 集群。在通过 TiDB Operator 在 [TiDB 集群监控](https://docs.pingcap.com/zh/tidb/stable/deploy-monitoring-services)中有一些监控系统配置的细节可供参考。 -在 v1.1 及更高版本的 TiDB Operator 中,可以通过简单的 CR 文件(即 TidbMonitor,可参考 [tidb-operator 中的示例](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/monitor/tidb-monitor.yaml))来快速建立对 Kubernetes 集群上的 TiDB 集群的监控。 +在 v1.1 及更高版本的 TiDB Operator 中,可以通过简单的 CR 文件(即 TidbMonitor,可参考 [tidb-operator 中的示例](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/monitor/tidb-monitor.yaml))来快速建立对 Kubernetes 集群上的 TiDB 集群的监控。 > **注意:** > @@ -82,13 +82,13 @@ basic-monitor Bound pvc-6db79253-cc9e-4730-bbba-ba987c29db6f 5G R 2. 设置 `spec.prometheus.config.configMapRef.name` 与 `spec.prometheus.config.configMapRef.namespace` 为自定义 ConfigMap 的名称与所属的 namespace。 3. 确认 TidbMonitor 是否已开启[动态配置功能](enable-monitor-dynamic-configuration.md),如果未开启该功能,需要重启 TidbMonitor 的 pod 重新加载配置。 -如需了解完整的配置示例,可参考 [tidb-operator 中的示例](https://github.com/pingcap/tidb-operator/blob/v1.5.0/examples/monitor-with-externalConfigMap/prometheus/README.md)。 +如需了解完整的配置示例,可参考 [tidb-operator 中的示例](https://github.com/pingcap/tidb-operator/blob/v1.5.1/examples/monitor-with-externalConfigMap/prometheus/README.md)。 #### 增加额外的命令行参数 设置 `spec.prometheus.config.commandOptions` 为用于启动 Prometheus 的额外的命令行参数。 -如需了解完整的配置示例,可参考 [tidb-operator 中的示例](https://github.com/pingcap/tidb-operator/blob/v1.5.0/examples/monitor-with-externalConfigMap/prometheus/README.md)。 +如需了解完整的配置示例,可参考 [tidb-operator 中的示例](https://github.com/pingcap/tidb-operator/blob/v1.5.1/examples/monitor-with-externalConfigMap/prometheus/README.md)。 > **注意:** > @@ -361,7 +361,7 @@ spec: imagePullPolicy: IfNotPresent ``` -如需了解完整的配置示例,可参考 TiDB Operator 仓库中的[示例](https://github.com/pingcap/tidb-operator/tree/v1.5.0/examples/monitor-multiple-cluster-non-tls)。 +如需了解完整的配置示例,可参考 TiDB Operator 仓库中的[示例](https://github.com/pingcap/tidb-operator/tree/v1.5.1/examples/monitor-multiple-cluster-non-tls)。 ### 使用 Grafana 查看多集群监控 diff --git a/zh/restore-from-aws-s3-by-snapshot.md b/zh/restore-from-aws-s3-by-snapshot.md index 104002232..ac5baf518 100644 --- a/zh/restore-from-aws-s3-by-snapshot.md +++ b/zh/restore-from-aws-s3-by-snapshot.md @@ -36,7 +36,7 @@ summary: 介绍如何将存储在 S3 上的备份元数据以及 EBS 卷快照 使用 TiDB Operator 将 S3 兼容存储上的备份元数据以及 EBS 快照恢复到 TiDB 之前,请按照以下步骤准备恢复环境。 -1. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml)。 +1. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml)。 2. 执行以下命令在 `test2` 这个命名空间中创建恢复需要的 RBAC 相关资源: diff --git a/zh/restore-from-aws-s3-using-br.md b/zh/restore-from-aws-s3-using-br.md index 9c1c43ffc..cc9d04512 100644 --- a/zh/restore-from-aws-s3-using-br.md +++ b/zh/restore-from-aws-s3-using-br.md @@ -39,7 +39,7 @@ PITR 全称为 Point-in-time recovery,该功能可以让你在新集群上恢 kubectl create namespace restore-test ``` -2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建恢复需要的 RBAC 相关资源: +2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建恢复需要的 RBAC 相关资源: {{< copyable "shell-regular" >}} @@ -247,7 +247,7 @@ demo2-restore-s3 Complete ... kubectl create namespace restore-test ``` -2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: +2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: ```shell kubectl apply -f backup-rbac.yaml -n restore-test diff --git a/zh/restore-from-azblob-using-br.md b/zh/restore-from-azblob-using-br.md index 43f1dab85..7c758ff78 100644 --- a/zh/restore-from-azblob-using-br.md +++ b/zh/restore-from-azblob-using-br.md @@ -38,7 +38,7 @@ PITR 全称为 Point-in-time recovery,该功能可以让你在新集群上恢 kubectl create namespace restore-test ``` -2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: +2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: ```shell kubectl apply -f backup-rbac.yaml -n restore-test @@ -155,7 +155,7 @@ demo2-restore-azblob Complete ... kubectl create namespace restore-test ``` -2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: +2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: ```shell kubectl apply -f backup-rbac.yaml -n restore-test diff --git a/zh/restore-from-gcs-using-br.md b/zh/restore-from-gcs-using-br.md index 5efb04e1d..f8d93fb31 100644 --- a/zh/restore-from-gcs-using-br.md +++ b/zh/restore-from-gcs-using-br.md @@ -39,7 +39,7 @@ PITR 全称为 Point-in-time recovery,该功能可以让你在新集群上恢 kubectl create namespace restore-test ``` -2. 下载文件 [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建恢复所需的 RBAC 相关资源: +2. 下载文件 [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建恢复所需的 RBAC 相关资源: {{< copyable "shell-regular" >}} @@ -161,7 +161,7 @@ PITR 全称为 Point-in-time recovery,该功能可以让你在新集群上恢 kubectl create namespace restore-test ``` -2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: +2. 下载文件 [backup-rbac.yaml](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `restore-test` 这个 namespace 中创建备份需要的 RBAC 相关资源: ```shell kubectl apply -f backup-rbac.yaml -n restore-test diff --git a/zh/restore-from-gcs.md b/zh/restore-from-gcs.md index 05b0342ca..33f15694c 100644 --- a/zh/restore-from-gcs.md +++ b/zh/restore-from-gcs.md @@ -28,7 +28,7 @@ TiDB Lightning 是一款将全量数据高速导入到 TiDB 集群的工具, ### 环境准备 -1. 下载文件 [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `test2` 这个 namespace 中创建恢复所需的 RBAC 相关资源: +1. 下载文件 [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `test2` 这个 namespace 中创建恢复所需的 RBAC 相关资源: {{< copyable "shell-regular" >}} diff --git a/zh/restore-from-pv-using-br.md b/zh/restore-from-pv-using-br.md index 4c494a011..110eb5e6a 100644 --- a/zh/restore-from-pv-using-br.md +++ b/zh/restore-from-pv-using-br.md @@ -22,7 +22,7 @@ summary: 介绍如何将存储在持久卷上的备份数据恢复到 TiDB 集 使用 BR 将 PV 上的备份数据恢复到 TiDB 前,请按照以下步骤准备恢复环境。 -1. 下载文件 [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml) 到执行恢复的服务器。 +1. 下载文件 [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml) 到执行恢复的服务器。 2. 执行以下命令在 `test2` 这个命名空间中创建恢复所需的 RBAC 相关资源: diff --git a/zh/restore-from-s3.md b/zh/restore-from-s3.md index 75fcd392a..b1f9c4003 100644 --- a/zh/restore-from-s3.md +++ b/zh/restore-from-s3.md @@ -28,7 +28,7 @@ TiDB Lightning 是一款将全量数据高速导入到 TiDB 集群的工具, ### 准备恢复环境 -1. 下载文件 [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.0/manifests/backup/backup-rbac.yaml),并执行以下命令在 `test2` 这个 namespace 中创建恢复所需的 RBAC 相关资源: +1. 下载文件 [`backup-rbac.yaml`](https://github.com/pingcap/tidb-operator/blob/v1.5.1/manifests/backup/backup-rbac.yaml),并执行以下命令在 `test2` 这个 namespace 中创建恢复所需的 RBAC 相关资源: {{< copyable "shell-regular" >}} diff --git a/zh/tidb-toolkit.md b/zh/tidb-toolkit.md index 118e72547..ba4e0576c 100644 --- a/zh/tidb-toolkit.md +++ b/zh/tidb-toolkit.md @@ -201,12 +201,12 @@ helm search repo pingcap ``` NAME CHART VERSION APP VERSION DESCRIPTION -pingcap/tidb-backup v1.5.0 A Helm chart for TiDB Backup or Restore -pingcap/tidb-cluster v1.5.0 A Helm chart for TiDB Cluster -pingcap/tidb-drainer v1.5.0 A Helm chart for TiDB Binlog drainer. -pingcap/tidb-lightning v1.5.0 A Helm chart for TiDB Lightning -pingcap/tidb-operator v1.5.0 v1.5.0 tidb-operator Helm chart for Kubernetes -pingcap/tikv-importer v1.5.0 A Helm chart for TiKV Importer +pingcap/tidb-backup v1.5.1 A Helm chart for TiDB Backup or Restore +pingcap/tidb-cluster v1.5.1 A Helm chart for TiDB Cluster +pingcap/tidb-drainer v1.5.1 A Helm chart for TiDB Binlog drainer. +pingcap/tidb-lightning v1.5.1 A Helm chart for TiDB Lightning +pingcap/tidb-operator v1.5.1 v1.5.1 tidb-operator Helm chart for Kubernetes +pingcap/tikv-importer v1.5.1 A Helm chart for TiKV Importer ``` 当新版本的 chart 发布后,你可以使用 `helm repo update` 命令更新本地对于仓库的缓存: @@ -266,9 +266,9 @@ helm uninstall ${release_name} -n ${namespace} {{< copyable "shell-regular" >}} ```shell -wget http://charts.pingcap.org/tidb-operator-v1.5.0.tgz -wget http://charts.pingcap.org/tidb-drainer-v1.5.0.tgz -wget http://charts.pingcap.org/tidb-lightning-v1.5.0.tgz +wget http://charts.pingcap.org/tidb-operator-v1.5.1.tgz +wget http://charts.pingcap.org/tidb-drainer-v1.5.1.tgz +wget http://charts.pingcap.org/tidb-lightning-v1.5.1.tgz ``` 将这些 chart 文件拷贝到服务器上并解压,可以通过 `helm install` 命令使用这些 chart 来安装相应组件,以 `tidb-operator` 为例: @@ -276,7 +276,7 @@ wget http://charts.pingcap.org/tidb-lightning-v1.5.0.tgz {{< copyable "shell-regular" >}} ```shell -tar zxvf tidb-operator.v1.5.0.tgz +tar zxvf tidb-operator.v1.5.1.tgz helm install ${release_name} ./tidb-operator --namespace=${namespace} ``` diff --git a/zh/upgrade-tidb-operator.md b/zh/upgrade-tidb-operator.md index 14ee1b091..9a57da2c1 100644 --- a/zh/upgrade-tidb-operator.md +++ b/zh/upgrade-tidb-operator.md @@ -70,27 +70,27 @@ aliases: ['/docs-cn/tidb-in-kubernetes/dev/upgrade-tidb-operator/'] kubectl get crd tidbclusters.pingcap.com ``` - 本文以 TiDB Operator v1.5.0 为例,你需要替换 `${operator_version}` 为你要升级到的 TiDB Operator 版本。 + 本文以 TiDB Operator v1.5.1 为例,你需要替换 `${operator_version}` 为你要升级到的 TiDB Operator 版本。 3. 获取你要升级的 `tidb-operator` chart 中的 `values.yaml` 文件: {{< copyable "shell-regular" >}} ```shell - mkdir -p ${HOME}/tidb-operator/v1.5.0 && \ - helm inspect values pingcap/tidb-operator --version=v1.5.0 > ${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml + mkdir -p ${HOME}/tidb-operator/v1.5.1 && \ + helm inspect values pingcap/tidb-operator --version=v1.5.1 > ${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml ``` -4. 修改 `${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml` 中 `operatorImage` 镜像版本为要升级到的版本。 +4. 修改 `${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml` 中 `operatorImage` 镜像版本为要升级到的版本。 -5. 如果你在旧版本 `values.yaml` 中设置了自定义配置,将自定义配置合并到 `${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml` 中。 +5. 如果你在旧版本 `values.yaml` 中设置了自定义配置,将自定义配置合并到 `${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml` 中。 6. 执行升级: {{< copyable "shell-regular" >}} ```shell - helm upgrade tidb-operator pingcap/tidb-operator --version=v1.5.0 -f ${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml -n tidb-admin + helm upgrade tidb-operator pingcap/tidb-operator --version=v1.5.1 -f ${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml -n tidb-admin ``` 7. Pod 全部正常启动之后,运行以下命令确认 TiDB Operator 镜像版本: @@ -101,13 +101,13 @@ aliases: ['/docs-cn/tidb-in-kubernetes/dev/upgrade-tidb-operator/'] kubectl get po -n tidb-admin -l app.kubernetes.io/instance=tidb-operator -o yaml | grep 'image:.*operator:' ``` - 如果输出类似下方的结果,则表示升级成功。其中,`v1.5.0` 表示已升级到的版本号。 + 如果输出类似下方的结果,则表示升级成功。其中,`v1.5.1` 表示已升级到的版本号。 ``` - image: pingcap/tidb-operator:v1.5.0 - image: docker.io/pingcap/tidb-operator:v1.5.0 - image: pingcap/tidb-operator:v1.5.0 - image: docker.io/pingcap/tidb-operator:v1.5.0 + image: pingcap/tidb-operator:v1.5.1 + image: docker.io/pingcap/tidb-operator:v1.5.1 + image: pingcap/tidb-operator:v1.5.1 + image: docker.io/pingcap/tidb-operator:v1.5.1 ``` > **注意:** @@ -138,14 +138,14 @@ aliases: ['/docs-cn/tidb-in-kubernetes/dev/upgrade-tidb-operator/'] wget -O crd.yaml https://raw.githubusercontent.com/pingcap/tidb-operator/${operator_version}/manifests/crd_v1beta1.yaml ``` - 本文以 TiDB Operator v1.5.0 为例,你需要替换 `${operator_version}` 为你要升级到的 TiDB Operator 版本。 + 本文以 TiDB Operator v1.5.1 为例,你需要替换 `${operator_version}` 为你要升级到的 TiDB Operator 版本。 2. 下载 `tidb-operator` chart 包文件: {{< copyable "shell-regular" >}} ```shell - wget http://charts.pingcap.org/tidb-operator-v1.5.0.tgz + wget http://charts.pingcap.org/tidb-operator-v1.5.1.tgz ``` 3. 下载 TiDB Operator 升级所需的 Docker 镜像: @@ -153,11 +153,11 @@ aliases: ['/docs-cn/tidb-in-kubernetes/dev/upgrade-tidb-operator/'] {{< copyable "shell-regular" >}} ```shell - docker pull pingcap/tidb-operator:v1.5.0 - docker pull pingcap/tidb-backup-manager:v1.5.0 + docker pull pingcap/tidb-operator:v1.5.1 + docker pull pingcap/tidb-backup-manager:v1.5.1 - docker save -o tidb-operator-v1.5.0.tar pingcap/tidb-operator:v1.5.0 - docker save -o tidb-backup-manager-v1.5.0.tar pingcap/tidb-backup-manager:v1.5.0 + docker save -o tidb-operator-v1.5.1.tar pingcap/tidb-operator:v1.5.1 + docker save -o tidb-backup-manager-v1.5.1.tar pingcap/tidb-backup-manager:v1.5.1 ``` 2. 将下载的文件和镜像上传到需要升级的服务器上,在服务器上按照以下步骤进行安装: @@ -185,9 +185,9 @@ aliases: ['/docs-cn/tidb-in-kubernetes/dev/upgrade-tidb-operator/'] {{< copyable "shell-regular" >}} ```shell - tar zxvf tidb-operator-v1.5.0.tgz && \ - mkdir -p ${HOME}/tidb-operator/v1.5.0 && \ - cp tidb-operator/values.yaml ${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml + tar zxvf tidb-operator-v1.5.1.tgz && \ + mkdir -p ${HOME}/tidb-operator/v1.5.1 && \ + cp tidb-operator/values.yaml ${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml ``` 4. 安装 Docker 镜像到服务器上: @@ -195,20 +195,20 @@ aliases: ['/docs-cn/tidb-in-kubernetes/dev/upgrade-tidb-operator/'] {{< copyable "shell-regular" >}} ```shell - docker load -i tidb-operator-v1.5.0.tar && \ - docker load -i tidb-backup-manager-v1.5.0.tar + docker load -i tidb-operator-v1.5.1.tar && \ + docker load -i tidb-backup-manager-v1.5.1.tar ``` -3. 修改 `${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml` 中 `operatorImage` 镜像版本为要升级到的版本。 +3. 修改 `${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml` 中 `operatorImage` 镜像版本为要升级到的版本。 -4. 如果你在旧版本 `values.yaml` 中设置了自定义配置,将自定义配置合并到 `${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml` 中。 +4. 如果你在旧版本 `values.yaml` 中设置了自定义配置,将自定义配置合并到 `${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml` 中。 5. 执行升级: {{< copyable "shell-regular" >}} ```shell - helm upgrade tidb-operator ./tidb-operator --version=v1.5.0 -f ${HOME}/tidb-operator/v1.5.0/values-tidb-operator.yaml + helm upgrade tidb-operator ./tidb-operator --version=v1.5.1 -f ${HOME}/tidb-operator/v1.5.1/values-tidb-operator.yaml ``` 6. Pod 全部正常启动之后,运行以下命令确认 TiDB Operator 镜像版本: @@ -219,13 +219,13 @@ aliases: ['/docs-cn/tidb-in-kubernetes/dev/upgrade-tidb-operator/'] kubectl get po -n tidb-admin -l app.kubernetes.io/instance=tidb-operator -o yaml | grep 'image:.*operator:' ``` - 如果输出类似下方的结果,则表示升级成功。其中,`v1.5.0` 表示已升级到的版本号。 + 如果输出类似下方的结果,则表示升级成功。其中,`v1.5.1` 表示已升级到的版本号。 ``` - image: pingcap/tidb-operator:v1.5.0 - image: docker.io/pingcap/tidb-operator:v1.5.0 - image: pingcap/tidb-operator:v1.5.0 - image: docker.io/pingcap/tidb-operator:v1.5.0 + image: pingcap/tidb-operator:v1.5.1 + image: docker.io/pingcap/tidb-operator:v1.5.1 + image: pingcap/tidb-operator:v1.5.1 + image: docker.io/pingcap/tidb-operator:v1.5.1 ``` > **注意:**