Skip to content

Console Output

Started by user Jenkins Admin
Obtained pipelines/pingcap/tiflow/latest/pull_cdc_integration_kafka_test.groovy from git https://github.com/PingCAP-QE/ci.git
Loading library tipipeline@main
Library tipipeline@main is cached. Copying from home.
[Pipeline] Start of Pipeline
[Pipeline] readJSON
[Pipeline] readTrusted
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-dtlq9-flzsw
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-dtlq9-flzsw is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-dtlq9-zmh9s
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "0af4cf79550a132f73b9269aab1c657b97df167a"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-dtlq9"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-dtlq9-flzsw"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-dtlq9-flzsw"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-dtlq9-flzsw"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-dtlq9-flzsw in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
[Pipeline] {
[Pipeline] stage
[Pipeline] { (Declarative: Checkout SCM)
[Pipeline] checkout
The recommended git tool is: git
No credentials specified
Cloning the remote Git repository
Using shallow clone with depth 1
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
 > git rev-list --no-walk 03312178c534dce949face80c69812d989e55009 # timeout=10
[Pipeline] }
[Pipeline] // stage
[Pipeline] withEnv
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 1 hr 5 min
[Pipeline] {
[Pipeline] stage
[Pipeline] { (Debug info)
[Pipeline] sh
+ printenv
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=4ee0bb83-541c-4330-a5b7-cf4c3f9cf92d
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-dtlq9-flzsw
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Debug info
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
_=/usr/bin/printenv
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
HUDSON_URL=https://do.pingcap.net/jenkins/
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=3
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-dtlq9
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-236df335481f9578f70eb859f68d5ceead3aa27f6c9385fda1ec4c08661c0305
NODE_LABELS=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-dtlq9 pingcap-tiflow-pull-cdc-integration-kafka-test-1845-dtlq9-flzsw
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-dtlq9-flzsw
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
+ echo -------------------------
-------------------------
+ go env
GO111MODULE=''
GOARCH='amd64'
GOBIN=''
GOCACHE='/home/jenkins/.cache/go-build'
GOENV='/home/jenkins/.config/go/env'
GOEXE=''
GOEXPERIMENT=''
GOFLAGS=''
GOHOSTARCH='amd64'
GOHOSTOS='linux'
GOINSECURE=''
GOMODCACHE='/go/pkg/mod'
GONOPROXY=''
GONOSUMDB=''
GOOS='linux'
GOPATH='/go'
GOPRIVATE=''
GOPROXY='http://goproxy.apps.svc,https://proxy.golang.org,direct'
GOROOT='/usr/local/go'
GOSUMDB='sum.golang.org'
GOTMPDIR=''
GOTOOLCHAIN='auto'
GOTOOLDIR='/usr/local/go/pkg/tool/linux_amd64'
GOVCS=''
GOVERSION='go1.21.0'
GCCGO='gccgo'
GOAMD64='v1'
AR='ar'
CC='gcc'
CXX='g++'
CGO_ENABLED='1'
GOMOD='/dev/null'
GOWORK=''
CGO_CFLAGS='-O2 -g'
CGO_CPPFLAGS=''
CGO_CXXFLAGS='-O2 -g'
CGO_FFLAGS='-O2 -g'
CGO_LDFLAGS='-O2 -g'
PKG_CONFIG='pkg-config'
GOGCCFLAGS='-fPIC -m64 -pthread -Wl,--no-gc-sections -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build1154451091=/tmp/go-build -gno-record-gcc-switches'
+ echo -------------------------
-------------------------
+ echo 'debug command: kubectl -n jenkins-tiflow exec -ti pingcap-tiflow-pull-cdc-integration-kafka-test-1845-dtlq9-flzsw bash'
debug command: kubectl -n jenkins-tiflow exec -ti pingcap-tiflow-pull-cdc-integration-kafka-test-1845-dtlq9-flzsw bash
[Pipeline] container
[Pipeline] {
[Pipeline] sh
+ dig github.com

; <<>> DiG 9.18.16 <<>> github.com
;; global options: +cmd
;; Got answer:
;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 63834
;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 1

;; OPT PSEUDOSECTION:
; EDNS: version: 0, flags:; udp: 1232
; COOKIE: 4832739f179ae496 (echoed)
;; QUESTION SECTION:
;github.com.			IN	A

;; ANSWER SECTION:
github.com.		16	IN	A	20.205.243.166

;; Query time: 1 msec
;; SERVER: 169.254.25.10#53(169.254.25.10) (UDP)
;; WHEN: Sat May 04 11:27:33 UTC 2024
;; MSG SIZE  rcvd: 77

[Pipeline] script
[Pipeline] {
[Pipeline] }
[Pipeline] // script
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Check diff files)
[Pipeline] container
[Pipeline] {
[Pipeline] script
[Pipeline] {
[Pipeline] withCredentials
Masking supported pattern matches of $token
[Pipeline] {
[Pipeline] httpRequest
Warning: A secret was passed to "httpRequest" using Groovy String interpolation, which is insecure.
		 Affected argument(s) used the following variable(s): [token]
		 See https://jenkins.io/redirect/groovy-string-interpolation for details.
HttpMethod: GET
URL: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=1&per_page=100
Content-Type: application/json
Authorization: *****
Sending request to url: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=1&per_page=100
Response Code: HTTP/1.1 200 OK
Success: Status code 200 is in the accepted range: 100:399
[Pipeline] httpRequest
Warning: A secret was passed to "httpRequest" using Groovy String interpolation, which is insecure.
		 Affected argument(s) used the following variable(s): [token]
		 See https://jenkins.io/redirect/groovy-string-interpolation for details.
HttpMethod: GET
URL: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=2&per_page=100
Content-Type: application/json
Authorization: *****
Sending request to url: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=2&per_page=100
Response Code: HTTP/1.1 200 OK
Success: Status code 200 is in the accepted range: 100:399
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] echo
pr_diff_files: [cdc/model/kv.go, cdc/model/sink.go, cdc/model/sink_test.go, cdc/processor/processor.go, cdc/processor/sinkmanager/manager.go, cdc/processor/sourcemanager/manager.go, cdc/redo/reader/reader.go, cdc/sink/dmlsink/factory/factory.go, cdc/sink/dmlsink/txn/mysql/dml.go, cdc/sink/dmlsink/txn/mysql/mysql.go, cdc/sink/dmlsink/txn/mysql/mysql_test.go, cmd/kafka-consumer/main.go, cmd/pulsar-consumer/main.go, cmd/storage-consumer/main.go, errors.toml, pkg/applier/redo.go, pkg/applier/redo_test.go, pkg/errors/cdc_errors.go, pkg/errors/helper.go, pkg/sink/codec/open/open_protocol_decoder.go, pkg/sink/codec/open/open_protocol_message.go, tests/integration_tests/_utils/check_sync_diff, tests/integration_tests/changefeed_dup_error_restart/conf/diff_config.toml, tests/integration_tests/changefeed_dup_error_restart/conf/workload, tests/integration_tests/changefeed_dup_error_restart/run.sh, tests/integration_tests/force_replicate_table/run.sh, tests/integration_tests/open_protocol_handle_key_only/data/data.sql, tests/integration_tests/open_protocol_handle_key_only/run.sh, tests/integration_tests/run_group.sh]
[Pipeline] echo
diff file not matched: cdc/model/kv.go
[Pipeline] }
[Pipeline] // script
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Checkout)
[Pipeline] timeout
Timeout set to expire in 10 min
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] cache
Cache restored successfully (git/pingcap/tiflow/rev-be15534-57a3c8e)
203865088 bytes in 1.61 secs (126393098 bytes/sec)
[Pipeline] {
[Pipeline] retry
[Pipeline] {
[Pipeline] script
[Pipeline] {
[Pipeline] sh
git version 2.36.6
Reinitialized existing Git repository in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/.git/
.git
HEAD is now at 57a3c8e82 f
POST git-upload-pack (656 bytes)
From https://github.com/pingcap/tiflow
 = [up to date]          master               -> origin/master
 = [up to date]          refs/pull/10919/head -> origin/pr/10919/head
Previous HEAD position was 57a3c8e82 f
HEAD is now at be1553484 codec(ticdc): avro simplify the unit test (#11010)
๐Ÿšง Checkouting to base SHA:be1553484fe4c03594eabb8d7435c694e5fd7224...
HEAD is now at be1553484 codec(ticdc): avro simplify the unit test (#11010)
โœ… Checked. ๐ŸŽ‰
๐Ÿงพ HEAD info:
be1553484fe4c03594eabb8d7435c694e5fd7224
be1553484 codec(ticdc): avro simplify the unit test (#11010)
2a7a65c6f Support Sequences (#10203)
36e9e1bf6 cli(ticdc): allow client authentication to be enabled without tls (#11005)
๐Ÿšง Pre-merge heads of pull requests to base SHA: be1553484fe4c03594eabb8d7435c694e5fd7224 ...
Updating be1553484..57a3c8e82
Fast-forward
 cdc/model/kv.go                                    |   5 +
 cdc/model/sink.go                                  |  38 ++-
 cdc/model/sink_test.go                             |   9 +-
 cdc/processor/processor.go                         |  21 +-
 cdc/processor/sinkmanager/manager.go               |   5 +
 cdc/processor/sourcemanager/manager.go             |  66 +++-
 cdc/redo/reader/reader.go                          |  21 +-
 cdc/sink/dmlsink/factory/factory.go                |   8 +-
 cdc/sink/dmlsink/txn/mysql/dml.go                  |  23 ++
 cdc/sink/dmlsink/txn/mysql/mysql.go                |  89 +++---
 cdc/sink/dmlsink/txn/mysql/mysql_test.go           |   2 +-
 cmd/kafka-consumer/main.go                         |   4 +-
 cmd/pulsar-consumer/main.go                        |  17 +-
 cmd/storage-consumer/main.go                       |   4 +-
 errors.toml                                        |   5 +
 pkg/applier/redo.go                                | 303 +++++++++++++++++-
 pkg/applier/redo_test.go                           | 347 ++++++++++++++++++++-
 pkg/errors/cdc_errors.go                           |   4 +
 pkg/errors/helper.go                               |  19 ++
 pkg/sink/codec/open/open_protocol_decoder.go       |   1 +
 pkg/sink/codec/open/open_protocol_message.go       |   3 +
 tests/integration_tests/_utils/check_sync_diff     |   2 +-
 .../conf/diff_config.toml                          |  29 ++
 .../changefeed_dup_error_restart/conf/workload     |  13 +
 .../changefeed_dup_error_restart/run.sh            |  54 ++++
 .../integration_tests/force_replicate_table/run.sh |   4 +-
 .../open_protocol_handle_key_only/data/data.sql    |   2 +-
 .../open_protocol_handle_key_only/run.sh           |   4 +-
 tests/integration_tests/run_group.sh               |   5 +-
 29 files changed, 1002 insertions(+), 105 deletions(-)
 create mode 100644 tests/integration_tests/changefeed_dup_error_restart/conf/diff_config.toml
 create mode 100644 tests/integration_tests/changefeed_dup_error_restart/conf/workload
 create mode 100755 tests/integration_tests/changefeed_dup_error_restart/run.sh
๐Ÿงพ Pre-merged result:
57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce
57a3c8e82 f
b2edc574f f
5bf93c6ca f
โœ… Pre merged ๐ŸŽ‰
โœ… ~~~~~All done.~~~~~~
[Pipeline] }
[Pipeline] // script
[Pipeline] }
[Pipeline] // retry
[Pipeline] }
Cache not saved (git/pingcap/tiflow/rev-be15534-57a3c8e already exists)
[Pipeline] // cache
[Pipeline] }
[Pipeline] // dir
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (prepare)
[Pipeline] timeout
Timeout set to expire in 20 min
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/third_party_download
[Pipeline] {
[Pipeline] retry
[Pipeline] {
[Pipeline] sh
+ cd ../tiflow
+ ./scripts/download-integration-test-binaries.sh master
Download binaries...
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0    590      0 --:--:-- --:--:-- --:--:--   594
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0   2573      0 --:--:-- --:--:-- --:--:--  2733
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0   1829      0 --:--:-- --:--:-- --:--:--  1863
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0   1745      0 --:--:-- --:--:-- --:--:--  1782
>>>
download tidb-server.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/tidb/600b2ed4bf0aa38224a1c4c4c68831820735515c/centos7/tidb-server.tar.gz
2024-05-04 19:27:53 URL:http://fileserver.pingcap.net/download/builds/pingcap/tidb/600b2ed4bf0aa38224a1c4c4c68831820735515c/centos7/tidb-server.tar.gz [536570515/536570515] -> "tmp/tidb-server.tar.gz" [1]
>>>
download pd-server.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/pd/1679dbca25b3483d1375c7e747da27e99ad77360/centos7/pd-server.tar.gz
2024-05-04 19:28:05 URL:http://fileserver.pingcap.net/download/builds/pingcap/pd/1679dbca25b3483d1375c7e747da27e99ad77360/centos7/pd-server.tar.gz [187372022/187372022] -> "tmp/pd-server.tar.gz" [1]
>>>
download tikv-server.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/tikv/72a0fd5b00235a7c56014b77ddd933e2a0d33c88/centos7/tikv-server.tar.gz
2024-05-04 19:28:25 URL:http://fileserver.pingcap.net/download/builds/pingcap/tikv/72a0fd5b00235a7c56014b77ddd933e2a0d33c88/centos7/tikv-server.tar.gz [919098782/919098782] -> "tmp/tikv-server.tar.gz" [1]
>>>
download tiflash.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/tiflash/master/8e170090fad91c94bef8d908e21c195c1d145b02/centos7/tiflash.tar.gz
2024-05-04 19:28:42 URL:http://fileserver.pingcap.net/download/builds/pingcap/tiflash/master/8e170090fad91c94bef8d908e21c195c1d145b02/centos7/tiflash.tar.gz [456057803/456057803] -> "tmp/tiflash.tar.gz" [1]
>>>
download minio.tar.gz from http://fileserver.pingcap.net/download/minio.tar.gz
2024-05-04 19:28:47 URL:http://fileserver.pingcap.net/download/minio.tar.gz [17718777/17718777] -> "tmp/minio.tar.gz" [1]
>>>
download go-ycsb from http://fileserver.pingcap.net/download/builds/pingcap/go-ycsb/test-br/go-ycsb
2024-05-04 19:28:49 URL:http://fileserver.pingcap.net/download/builds/pingcap/go-ycsb/test-br/go-ycsb [45975512/45975512] -> "third_bin/go-ycsb" [1]
>>>
download jq from http://fileserver.pingcap.net/download/builds/pingcap/test/jq-1.6/jq-linux64
2024-05-04 19:28:49 URL:http://fileserver.pingcap.net/download/builds/pingcap/test/jq-1.6/jq-linux64 [3953824/3953824] -> "third_bin/jq" [1]
>>>
download etcd.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/cdc/etcd-v3.4.7-linux-amd64.tar.gz
2024-05-04 19:28:49 URL:http://fileserver.pingcap.net/download/builds/pingcap/cdc/etcd-v3.4.7-linux-amd64.tar.gz [17310840/17310840] -> "tmp/etcd.tar.gz" [1]
>>>
download sync_diff_inspector.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/cdc/sync_diff_inspector_hash-d671b084_linux-amd64.tar.gz
2024-05-04 19:28:52 URL:http://fileserver.pingcap.net/download/builds/pingcap/cdc/sync_diff_inspector_hash-d671b084_linux-amd64.tar.gz [79877126/79877126] -> "tmp/sync_diff_inspector.tar.gz" [1]
>>>
download schema-registry.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/cdc/schema-registry.tar.gz
2024-05-04 19:29:00 URL:http://fileserver.pingcap.net/download/builds/pingcap/cdc/schema-registry.tar.gz [278386006/278386006] -> "tmp/schema-registry.tar.gz" [1]
Download SUCCESS
+ ls -alh ./bin
total 1.9G
drwxr-sr-x  6 jenkins jenkins  4.0K May  4 19:29 .
drwxr-sr-x 19 jenkins jenkins  4.0K May  4 19:29 ..
drwxr-sr-x  2 jenkins jenkins  4.0K May 19  2023 bin
drwxr-sr-x  4 jenkins jenkins  4.0K May 10  2023 etc
-rwxr-xr-x  1 jenkins jenkins   17M Apr  2  2020 etcdctl
-rwxr-xr-x  1 jenkins jenkins   44M May  4 19:28 go-ycsb
-rwxr-xr-x  1 jenkins jenkins  3.8M May  4 19:28 jq
drwxr-sr-x  3 jenkins jenkins  4.0K May 10  2023 lib
lrwxrwxrwx  1 jenkins jenkins    13 Apr 30 11:15 libc++.so.1 -> libc++.so.1.0
-rwxr-xr-x  1 jenkins jenkins 1016K Nov  7 01:00 libc++.so.1.0
lrwxrwxrwx  1 jenkins jenkins    16 Apr 30 11:15 libc++abi.so.1 -> libc++abi.so.1.0
-rwxr-xr-x  1 jenkins jenkins  358K Nov  7 01:00 libc++abi.so.1.0
lrwxrwxrwx  1 jenkins jenkins    13 Apr 30 11:15 libgmssl.so -> libgmssl.so.3
lrwxrwxrwx  1 jenkins jenkins    15 Apr 30 11:15 libgmssl.so.3 -> libgmssl.so.3.0
-rwxr-xr-x  1 jenkins jenkins  2.6M Apr 30 10:34 libgmssl.so.3.0
-rwxr-xr-x  1 jenkins jenkins  272M Apr 30 11:16 libtiflash_proxy.so
-rwxr-xr-x  1 jenkins jenkins   50M Jul 29  2020 minio
-rwxr-xr-x  1 jenkins jenkins   37M Apr 30 16:11 pd-api-bench
-rwxr-xr-x  1 jenkins jenkins   44M Apr 30 16:10 pd-ctl
-rwxr-xr-x  1 jenkins jenkins   36M Apr 30 16:10 pd-heartbeat-bench
-rwxr-xr-x  1 jenkins jenkins   32M Apr 30 16:10 pd-recover
-rwxr-xr-x  1 jenkins jenkins  106M Apr 30 16:10 pd-server
-rwxr-xr-x  1 jenkins jenkins   26M Apr 30 16:10 pd-tso-bench
-rwxr-xr-x  1 jenkins jenkins  3.0M Apr 30 16:11 pd-ut
-rwxr-xr-x  1 jenkins jenkins   32M Apr 30 16:10 regions-dump
drwxr-sr-x  4 jenkins jenkins  4.0K May 10  2023 share
-rwxr-xr-x  1 jenkins jenkins   32M Apr 30 16:11 stores-dump
-rwxr-xr-x  1 jenkins jenkins  192M Sep 22  2023 sync_diff_inspector
-rwxr-xr-x  1 jenkins jenkins  208M May  1 10:57 tidb-server
-rwxr-xr-x  1 jenkins jenkins  380M Apr 30 11:15 tiflash
-rwxr-xr-x  1 jenkins jenkins  418M Apr 30 11:29 tikv-server
-rwxr-xr-x  1 jenkins jenkins  2.0M Apr 30 16:11 xprog
+ make check_third_party_binary
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/tidb-server
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/tikv-server
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/pd-server
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/tiflash
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/pd-ctl
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/sync_diff_inspector
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/go-ycsb
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/etcdctl
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/jq
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/minio
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/bin/schema-registry-start
+ cd -
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/third_party_download
+ mkdir -p bin
+ mv ../tiflow/bin/bin ../tiflow/bin/etc ../tiflow/bin/etcdctl ../tiflow/bin/go-ycsb ../tiflow/bin/jq ../tiflow/bin/lib ../tiflow/bin/libc++.so.1 ../tiflow/bin/libc++.so.1.0 ../tiflow/bin/libc++abi.so.1 ../tiflow/bin/libc++abi.so.1.0 ../tiflow/bin/libgmssl.so ../tiflow/bin/libgmssl.so.3 ../tiflow/bin/libgmssl.so.3.0 ../tiflow/bin/libtiflash_proxy.so ../tiflow/bin/minio ../tiflow/bin/pd-api-bench ../tiflow/bin/pd-ctl ../tiflow/bin/pd-heartbeat-bench ../tiflow/bin/pd-recover ../tiflow/bin/pd-server ../tiflow/bin/pd-tso-bench ../tiflow/bin/pd-ut ../tiflow/bin/regions-dump ../tiflow/bin/share ../tiflow/bin/stores-dump ../tiflow/bin/sync_diff_inspector ../tiflow/bin/tidb-server ../tiflow/bin/tiflash ../tiflow/bin/tikv-server ../tiflow/bin/xprog ./bin/
+ ls -alh ./bin
total 1.9G
drwxr-sr-x 6 jenkins jenkins  4.0K May  4 19:29 .
drwxr-sr-x 3 jenkins jenkins  4.0K May  4 19:29 ..
drwxr-sr-x 2 jenkins jenkins  4.0K May 19  2023 bin
drwxr-sr-x 4 jenkins jenkins  4.0K May 10  2023 etc
-rwxr-xr-x 1 jenkins jenkins   17M Apr  2  2020 etcdctl
-rwxr-xr-x 1 jenkins jenkins   44M May  4 19:28 go-ycsb
-rwxr-xr-x 1 jenkins jenkins  3.8M May  4 19:28 jq
drwxr-sr-x 3 jenkins jenkins  4.0K May 10  2023 lib
lrwxrwxrwx 1 jenkins jenkins    13 Apr 30 11:15 libc++.so.1 -> libc++.so.1.0
-rwxr-xr-x 1 jenkins jenkins 1016K Nov  7 01:00 libc++.so.1.0
lrwxrwxrwx 1 jenkins jenkins    16 Apr 30 11:15 libc++abi.so.1 -> libc++abi.so.1.0
-rwxr-xr-x 1 jenkins jenkins  358K Nov  7 01:00 libc++abi.so.1.0
lrwxrwxrwx 1 jenkins jenkins    13 Apr 30 11:15 libgmssl.so -> libgmssl.so.3
lrwxrwxrwx 1 jenkins jenkins    15 Apr 30 11:15 libgmssl.so.3 -> libgmssl.so.3.0
-rwxr-xr-x 1 jenkins jenkins  2.6M Apr 30 10:34 libgmssl.so.3.0
-rwxr-xr-x 1 jenkins jenkins  272M Apr 30 11:16 libtiflash_proxy.so
-rwxr-xr-x 1 jenkins jenkins   50M Jul 29  2020 minio
-rwxr-xr-x 1 jenkins jenkins   37M Apr 30 16:11 pd-api-bench
-rwxr-xr-x 1 jenkins jenkins   44M Apr 30 16:10 pd-ctl
-rwxr-xr-x 1 jenkins jenkins   36M Apr 30 16:10 pd-heartbeat-bench
-rwxr-xr-x 1 jenkins jenkins   32M Apr 30 16:10 pd-recover
-rwxr-xr-x 1 jenkins jenkins  106M Apr 30 16:10 pd-server
-rwxr-xr-x 1 jenkins jenkins   26M Apr 30 16:10 pd-tso-bench
-rwxr-xr-x 1 jenkins jenkins  3.0M Apr 30 16:11 pd-ut
-rwxr-xr-x 1 jenkins jenkins   32M Apr 30 16:10 regions-dump
drwxr-sr-x 4 jenkins jenkins  4.0K May 10  2023 share
-rwxr-xr-x 1 jenkins jenkins   32M Apr 30 16:11 stores-dump
-rwxr-xr-x 1 jenkins jenkins  192M Sep 22  2023 sync_diff_inspector
-rwxr-xr-x 1 jenkins jenkins  208M May  1 10:57 tidb-server
-rwxr-xr-x 1 jenkins jenkins  380M Apr 30 11:15 tiflash
-rwxr-xr-x 1 jenkins jenkins  418M Apr 30 11:29 tikv-server
-rwxr-xr-x 1 jenkins jenkins  2.0M Apr 30 16:11 xprog
+ ./bin/tidb-server -V
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
+ ./bin/pd-server -V
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
+ ./bin/tikv-server -V
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
+ ./bin/tiflash --version
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
+ ./bin/sync_diff_inspector --version
App Name: sync_diff_inspector v2.0
Release Version: v7.4.0
Git Commit Hash: d671b0840063bc2532941f02e02e12627402844c
Git Branch: heads/refs/tags/v7.4.0
UTC Build Time: 2023-09-22 03:51:56
Go Version: go1.21.1
[Pipeline] }
[Pipeline] // retry
[Pipeline] }
[Pipeline] // dir
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] cache
Cache not restored (no such key found)
[Pipeline] {
[Pipeline] sh
+ ls -alh ./bin
total 8.0K
drwxr-sr-x  2 jenkins jenkins 4.0K May  4 19:29 .
drwxr-sr-x 19 jenkins jenkins 4.0K May  4 19:29 ..
+ '[' -f ./bin/cdc ']'
+ make cdc
CGO_ENABLED=0 GO111MODULE=on go build  -trimpath  -ldflags '-X "github.com/pingcap/tiflow/pkg/version.ReleaseVersion=v8.2.0-alpha-88-g57a3c8e82" -X "github.com/pingcap/tiflow/pkg/version.BuildTS=2024-05-04 11:29:04" -X "github.com/pingcap/tiflow/pkg/version.GitHash=57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce" -X "github.com/pingcap/tiflow/pkg/version.GitBranch=HEAD" -X "github.com/pingcap/tiflow/pkg/version.GoVersion=go version go1.21.0 linux/amd64" -X "github.com/pingcap/tidb/pkg/parser/mysql.TiDBReleaseVersion=v8.2.0-alpha-88-g57a3c8e82"' -o bin/cdc ./cmd/cdc
go: downloading github.com/pingcap/kvproto v0.0.0-20240227073058-929ab83f9754
go: downloading github.com/pingcap/failpoint v0.0.0-20220801062533-2eaa32854a6c
go: downloading github.com/spf13/cobra v1.8.0
go: downloading github.com/pingcap/log v1.1.1-0.20240314023424-862ccc32f18d
go: downloading github.com/pingcap/sysutil v1.0.1-0.20240311050922-ae81ee01f3a5
go: downloading go.uber.org/zap v1.27.0
go: downloading github.com/BurntSushi/toml v1.3.2
go: downloading github.com/pingcap/tidb v1.1.0-beta.0.20240415145106-cd9c676e9ba4
go: downloading github.com/pingcap/errors v0.11.5-0.20240318064555-6bd07397691f
go: downloading github.com/pingcap/tidb/pkg/parser v0.0.0-20240410110152-5fc42c9be2f5
go: downloading github.com/prometheus/client_golang v1.19.0
go: downloading github.com/tinylib/msgp v1.1.6
go: downloading golang.org/x/sync v0.7.0
go: downloading google.golang.org/grpc v1.62.1
go: downloading golang.org/x/net v0.24.0
go: downloading github.com/spf13/pflag v1.0.5
go: downloading github.com/gin-gonic/gin v1.9.1
go: downloading github.com/dustin/go-humanize v1.0.1
go: downloading github.com/fatih/color v1.16.0
go: downloading github.com/tikv/client-go/v2 v2.0.8-0.20240409022718-714958ccd4d5
go: downloading github.com/tikv/pd/client v0.0.0-20240322051414-fb9e2d561b6e
go: downloading github.com/coreos/go-semver v0.3.1
go: downloading github.com/IBM/sarama v1.41.2
go: downloading github.com/xdg/scram v1.0.5
go: downloading github.com/DATA-DOG/go-sqlmock v1.5.0
go: downloading github.com/go-sql-driver/mysql v1.7.1
go: downloading github.com/imdario/mergo v0.3.16
go: downloading go.uber.org/multierr v1.11.0
go: downloading github.com/swaggo/gin-swagger v1.2.0
go: downloading github.com/swaggo/files v0.0.0-20210815190702-a29dd2bc99b2
go: downloading github.com/google/btree v1.1.2
go: downloading github.com/google/uuid v1.6.0
go: downloading go.etcd.io/etcd/client/v3 v3.5.12
go: downloading go.etcd.io/etcd/server/v3 v3.5.12
go: downloading golang.org/x/time v0.5.0
go: downloading github.com/apache/pulsar-client-go v0.11.0
go: downloading github.com/aws/aws-sdk-go-v2 v1.19.1
go: downloading github.com/pingcap/tidb-tools v0.0.0-20240305021104-9f9bea84490b
go: downloading go.uber.org/atomic v1.11.0
go: downloading github.com/soheilhy/cmux v0.1.5
go: downloading cloud.google.com/go/storage v1.39.1
go: downloading github.com/Azure/azure-sdk-for-go/sdk/azcore v1.9.1
go: downloading github.com/KimMachineGun/automemlimit v0.2.4
go: downloading github.com/aws/aws-sdk-go v1.50.0
go: downloading github.com/tikv/pd v1.1.0-beta.0.20240407022249-7179657d129b
go: downloading github.com/json-iterator/go v1.1.12
go: downloading github.com/shirou/gopsutil/v3 v3.24.2
go: downloading github.com/stretchr/testify v1.9.0
go: downloading github.com/gogo/protobuf v1.3.2
go: downloading cloud.google.com/go v0.112.2
go: downloading github.com/cenkalti/backoff/v4 v4.2.1
go: downloading github.com/goccy/go-json v0.10.2
go: downloading go.etcd.io/etcd/api/v3 v3.5.12
go: downloading github.com/cockroachdb/pebble v1.1.0
go: downloading github.com/benbjohnson/clock v1.3.5
go: downloading go.etcd.io/etcd/client/pkg/v3 v3.5.12
go: downloading github.com/grpc-ecosystem/go-grpc-prometheus v1.2.0
go: downloading github.com/modern-go/reflect2 v1.0.2
go: downloading github.com/phayes/freeport v0.0.0-20180830031419-95f893ade6f2
go: downloading golang.org/x/sys v0.19.0
go: downloading github.com/r3labs/diff v1.1.0
go: downloading github.com/YangKeao/seahash v0.0.0-20240229041150-e7bf269c3140
go: downloading github.com/mattn/go-colorable v0.1.13
go: downloading github.com/mattn/go-isatty v0.0.20
go: downloading github.com/klauspost/compress v1.17.8
go: downloading github.com/pierrec/lz4/v4 v4.1.18
go: downloading github.com/uber-go/atomic v1.4.0
go: downloading go.etcd.io/etcd/pkg/v3 v3.5.12
go: downloading github.com/go-mysql-org/go-mysql v1.7.1-0.20240314115043-2199dfb0ba98
go: downloading github.com/pingcap/check v0.0.0-20211026125417-57bd13f7b5f0
go: downloading github.com/robfig/cron v1.2.0
go: downloading gopkg.in/natefinch/lumberjack.v2 v2.2.1
go: downloading github.com/jcmturner/gokrb5/v8 v8.4.4
go: downloading github.com/segmentio/kafka-go v0.4.41-0.20230526171612-f057b1d369cd
go: downloading github.com/gavv/monotime v0.0.0-20190418164738-30dba4353424
go: downloading github.com/swaggo/swag v1.16.3
go: downloading github.com/rcrowley/go-metrics v0.0.0-20201227073835-cf1acfcdf475
go: downloading golang.org/x/oauth2 v0.18.0
go: downloading github.com/hashicorp/golang-lru v0.5.1
go: downloading github.com/pingcap/tidb-dashboard v0.0.0-20240326110213-9768844ff5d7
go: downloading github.com/golang/mock v1.6.0
go: downloading github.com/xdg/stringprep v1.0.3
go: downloading golang.org/x/crypto v0.22.0
go: downloading gorm.io/gorm v1.24.5
go: downloading github.com/philhofer/fwd v1.1.1
go: downloading github.com/containerd/cgroups v1.0.4
go: downloading github.com/gin-contrib/sse v0.1.0
go: downloading github.com/go-playground/validator/v10 v10.14.0
go: downloading github.com/pelletier/go-toml/v2 v2.0.8
go: downloading github.com/ugorji/go/codec v1.2.11
go: downloading google.golang.org/protobuf v1.33.0
go: downloading gopkg.in/yaml.v3 v3.0.1
go: downloading github.com/prometheus/client_model v0.6.1
go: downloading github.com/prometheus/common v0.52.2
go: downloading github.com/beorn7/perks v1.0.1
go: downloading github.com/cespare/xxhash/v2 v2.3.0
go: downloading github.com/prometheus/procfs v0.13.0
go: downloading github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd
go: downloading github.com/aws/smithy-go v1.13.5
go: downloading github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc
go: downloading github.com/eapache/go-resiliency v1.4.0
go: downloading github.com/eapache/go-xerial-snappy v0.0.0-20230731223053-c322873962e3
go: downloading github.com/eapache/queue v1.1.0
go: downloading github.com/hashicorp/go-multierror v1.1.1
go: downloading github.com/jcmturner/gofork v1.7.6
go: downloading github.com/golang/protobuf v1.5.4
go: downloading github.com/opentracing/opentracing-go v1.2.0
go: downloading github.com/pkg/errors v0.9.1
go: downloading github.com/tiancaiamao/gp v0.0.0-20221230034425-4025bc8a4d4a
go: downloading github.com/bits-and-blooms/bitset v1.4.0
go: downloading github.com/linkedin/goavro/v2 v2.11.1
go: downloading github.com/sirupsen/logrus v1.9.3
go: downloading google.golang.org/genproto/googleapis/api v0.0.0-20240401170217-c3f982113cda
go: downloading google.golang.org/genproto v0.0.0-20240401170217-c3f982113cda
go: downloading cloud.google.com/go/compute/metadata v0.2.3
go: downloading cloud.google.com/go/iam v1.1.7
go: downloading cloud.google.com/go/compute v1.25.1
go: downloading github.com/googleapis/gax-go/v2 v2.12.3
go: downloading google.golang.org/api v0.170.0
go: downloading github.com/pierrec/lz4 v2.6.1+incompatible
go: downloading github.com/grpc-ecosystem/grpc-gateway v1.16.0
go: downloading go.etcd.io/bbolt v1.3.9
go: downloading github.com/tmc/grpc-websocket-proxy v0.0.0-20220101234140-673ab2c3ae75
go: downloading go.opentelemetry.io/contrib/instrumentation/google.golang.org/grpc/otelgrpc v0.49.0
go: downloading go.opentelemetry.io/otel v1.24.0
go: downloading go.opentelemetry.io/otel/sdk v1.22.0
go: downloading go.opentelemetry.io/otel/exporters/otlp/otlptrace/otlptracegrpc v1.22.0
go: downloading sigs.k8s.io/yaml v1.4.0
go: downloading github.com/coreos/go-systemd/v22 v22.5.0
go: downloading github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2
go: downloading github.com/stretchr/objx v0.5.2
go: downloading golang.org/x/term v0.19.0
go: downloading github.com/tklauser/go-sysconf v0.3.12
go: downloading github.com/cakturk/go-netstat v0.0.0-20200220111822-e5b49efee7a5
go: downloading google.golang.org/genproto/googleapis/rpc v0.0.0-20240401170217-c3f982113cda
go: downloading github.com/docker/go-units v0.5.0
go: downloading github.com/Masterminds/semver v1.5.0
go: downloading github.com/siddontang/go-log v0.0.0-20180807004314-8d05993dda07
go: downloading github.com/siddontang/go v0.0.0-20180604090527-bdc77568d726
go: downloading github.com/hashicorp/go-uuid v1.0.3
go: downloading github.com/AthenZ/athenz v1.10.39
go: downloading github.com/jcmturner/dnsutils/v2 v2.0.0
go: downloading github.com/xdg-go/scram v1.1.2
go: downloading golang.org/x/text v0.14.0
go: downloading github.com/go-openapi/spec v0.21.0
go: downloading golang.org/x/tools v0.20.0
go: downloading github.com/KyleBanks/depth v1.2.1
go: downloading github.com/jinzhu/now v1.1.5
go: downloading github.com/joomcode/errorx v1.0.1
go: downloading github.com/cockroachdb/errors v1.11.1
go: downloading github.com/cockroachdb/redact v1.1.5
go: downloading github.com/cockroachdb/tokenbucket v0.0.0-20230807174530-cc333fc44b06
go: downloading golang.org/x/exp v0.0.0-20240409090435-93d18d7e34b8
go: downloading github.com/glebarez/sqlite v1.7.0
go: downloading gorm.io/driver/mysql v1.3.3
go: downloading github.com/godbus/dbus/v5 v5.0.4
go: downloading github.com/opencontainers/runtime-spec v1.0.2
go: downloading github.com/cilium/ebpf v0.4.0
go: downloading github.com/gabriel-vasile/mimetype v1.4.2
go: downloading github.com/go-playground/universal-translator v0.18.1
go: downloading github.com/leodido/go-urn v1.2.4
go: downloading github.com/Azure/azure-sdk-for-go/sdk/internal v1.5.1
go: downloading go.opentelemetry.io/otel/exporters/otlp/otlptrace v1.22.0
go: downloading github.com/golang/snappy v0.0.4
go: downloading github.com/hashicorp/errwrap v1.0.0
go: downloading github.com/grpc-ecosystem/go-grpc-middleware v1.4.0
go: downloading github.com/cznic/mathutil v0.0.0-20181122101859-297441e03548
go: downloading github.com/twmb/murmur3 v1.1.6
go: downloading github.com/dgryski/go-farm v0.0.0-20200201041132-a6ae2369ad13
go: downloading github.com/cloudfoundry/gosigar v1.3.6
go: downloading github.com/spaolacci/murmur3 v1.1.0
go: downloading golang.org/x/mod v0.17.0
go: downloading go.opencensus.io v0.23.1-0.20220331163232-052120675fac
go: downloading github.com/godbus/dbus v0.0.0-20190726142602-4481cbc300e2
go: downloading go.opentelemetry.io/otel/trace v1.24.0
go: downloading github.com/uber/jaeger-client-go v2.30.0+incompatible
go: downloading go.etcd.io/etcd/raft/v3 v3.5.12
go: downloading github.com/xiang90/probing v0.0.0-20221125231312-a49e3df8f510
go: downloading github.com/Azure/azure-sdk-for-go/sdk/storage/azblob v1.0.0
go: downloading github.com/Azure/azure-sdk-for-go/sdk/azidentity v1.5.1
go: downloading github.com/aliyun/alibaba-cloud-sdk-go v1.61.1581
go: downloading github.com/jonboulle/clockwork v0.4.0
go: downloading github.com/go-resty/resty/v2 v2.11.0
go: downloading github.com/ks3sdklib/aws-sdk-go v1.2.9
go: downloading github.com/coocood/freecache v1.2.1
go: downloading github.com/pingcap/tipb v0.0.0-20240318032315-55a7867ddd50
go: downloading github.com/ngaut/pools v0.0.0-20180318154953-b7bc8c42aac7
go: downloading github.com/otiai10/copy v1.2.0
go: downloading github.com/gorilla/mux v1.8.0
go: downloading github.com/blacktear23/go-proxyprotocol v1.0.6
go: downloading github.com/pingcap/fn v1.0.0
go: downloading github.com/tiancaiamao/appdash v0.0.0-20181126055449-889f96f722a2
go: downloading github.com/sourcegraph/appdash-data v0.0.0-20151005221446-73f23eafcf67
go: downloading github.com/scalalang2/golang-fifo v0.1.5
go: downloading github.com/shopspring/decimal v1.3.0
go: downloading github.com/tidwall/btree v1.7.0
go: downloading github.com/pingcap/goleveldb v0.0.0-20191226122134-f82aafb29989
go: downloading github.com/jcmturner/aescts/v2 v2.0.0
go: downloading github.com/jcmturner/rpc/v2 v2.0.3
go: downloading github.com/99designs/keyring v1.2.1
go: downloading github.com/golang-jwt/jwt v3.2.2+incompatible
go: downloading github.com/aws/aws-sdk-go-v2/config v1.18.30
go: downloading github.com/aws/aws-sdk-go-v2/credentials v1.13.29
go: downloading github.com/aws/aws-sdk-go-v2/service/glue v1.58.1
go: downloading github.com/jarcoal/httpmock v1.2.0
go: downloading github.com/mailru/easyjson v0.7.7
go: downloading github.com/jinzhu/inflection v1.0.0
go: downloading github.com/glebarez/go-sqlite v1.21.2
go: downloading github.com/go-ozzo/ozzo-validation/v4 v4.3.0
go: downloading github.com/gorilla/websocket v1.5.1
go: downloading go.opentelemetry.io/otel/metric v1.24.0
go: downloading github.com/xdg-go/pbkdf2 v1.0.0
go: downloading github.com/xdg-go/stringprep v1.0.4
go: downloading github.com/go-playground/locales v0.14.1
go: downloading github.com/tklauser/numcpus v0.6.1
go: downloading github.com/edwingeng/deque v0.0.0-20191220032131-8596380dee17
go: downloading github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec
go: downloading github.com/golang-jwt/jwt/v4 v4.5.0
go: downloading github.com/cockroachdb/logtags v0.0.0-20230118201751-21c54148d20b
go: downloading github.com/getsentry/sentry-go v0.27.0
go: downloading github.com/go-openapi/jsonpointer v0.21.0
go: downloading github.com/go-openapi/jsonreference v0.21.0
go: downloading github.com/go-openapi/swag v0.23.0
go: downloading github.com/AzureAD/microsoft-authentication-library-for-go v1.2.1
go: downloading github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da
go: downloading github.com/uber/jaeger-lib v2.4.1+incompatible
go: downloading github.com/go-logr/logr v1.4.1
go: downloading go.etcd.io/etcd/client/v2 v2.305.12
go: downloading github.com/jellydator/ttlcache/v3 v3.0.1
go: downloading gopkg.in/yaml.v2 v2.4.0
go: downloading go.uber.org/mock v0.4.0
go: downloading go.opentelemetry.io/proto/otlp v1.1.0
go: downloading github.com/jfcg/sorty/v2 v2.1.0
go: downloading github.com/carlmjohnson/flagext v0.21.0
go: downloading github.com/dolthub/swiss v0.2.1
go: downloading github.com/dgraph-io/ristretto v0.1.1
go: downloading github.com/lestrrat-go/jwx/v2 v2.0.21
go: downloading github.com/influxdata/tdigest v0.0.1
go: downloading github.com/ngaut/sync2 v0.0.0-20141008032647-7a24ed77b2ef
go: downloading github.com/danjacques/gofslock v0.0.0-20240212154529-d899e02bfe22
go: downloading github.com/opentracing/basictracer-go v1.1.0
go: downloading github.com/sourcegraph/appdash v0.0.0-20190731080439-ebfcffb1b5c0
go: downloading github.com/asaskevich/govalidator v0.0.0-20230301143203-a9d515a09cc2
go: downloading github.com/cheggaaa/pb/v3 v3.0.8
go: downloading github.com/yangkeao/ldap/v3 v3.4.5-0.20230421065457-369a3bab1117
go: downloading github.com/google/pprof v0.0.0-20240117000934-35fc243c5815
go: downloading github.com/dvsekhvalnov/jose2go v1.5.0
go: downloading github.com/gsterjov/go-libsecret v0.0.0-20161001094733-a6f4afe4910c
go: downloading github.com/mtibben/percent v0.2.1
go: downloading github.com/spkg/bom v1.0.0
go: downloading github.com/xitongsys/parquet-go v1.6.0
go: downloading github.com/aws/aws-sdk-go-v2/feature/ec2/imds v1.13.6
go: downloading github.com/aws/aws-sdk-go-v2/internal/ini v1.3.37
go: downloading github.com/aws/aws-sdk-go-v2/service/sso v1.12.14
go: downloading github.com/aws/aws-sdk-go-v2/service/ssooidc v1.14.14
go: downloading github.com/aws/aws-sdk-go-v2/service/sts v1.20.1
go: downloading github.com/grpc-ecosystem/grpc-gateway/v2 v2.19.1
go: downloading modernc.org/libc v1.37.1
go: downloading modernc.org/sqlite v1.27.0
go: downloading github.com/kr/pretty v0.3.1
go: downloading github.com/go-logr/stdr v1.2.2
go: downloading github.com/joho/sqltocsv v0.0.0-20210428211105-a6d6801d59df
go: downloading github.com/jedib0t/go-pretty/v6 v6.2.2
go: downloading github.com/jfcg/sixb v1.3.8
go: downloading github.com/wangjohn/quickselect v0.0.0-20161129230411-ed8402a42d5f
go: downloading github.com/kylelemons/godebug v1.1.0
go: downloading github.com/pkg/browser v0.0.0-20240102092130-5ac0b6a4141c
go: downloading github.com/robfig/cron/v3 v3.0.1
go: downloading github.com/dolthub/maphash v0.1.0
go: downloading github.com/vbauerster/mpb/v7 v7.5.3
go: downloading github.com/aws/aws-sdk-go-v2/internal/configsources v1.1.36
go: downloading github.com/pingcap/badger v1.5.1-0.20230103063557-828f39b09b6d
go: downloading github.com/VividCortex/ewma v1.2.0
go: downloading github.com/mattn/go-runewidth v0.0.15
go: downloading github.com/Azure/go-ntlmssp v0.0.0-20221128193559-754e69321358
go: downloading github.com/go-asn1-ber/asn1-ber v1.5.4
go: downloading github.com/lestrrat-go/blackmagic v1.0.2
go: downloading github.com/lestrrat-go/httprc v1.0.5
go: downloading github.com/lestrrat-go/iter v1.0.2
go: downloading github.com/lestrrat-go/option v1.0.1
go: downloading github.com/apache/thrift v0.16.0
go: downloading github.com/josharian/intern v1.0.0
go: downloading github.com/aws/aws-sdk-go-v2/service/internal/presigned-url v1.9.30
go: downloading github.com/kr/text v0.2.0
go: downloading github.com/rogpeppe/go-internal v1.12.0
go: downloading github.com/golang-jwt/jwt/v5 v5.2.0
go: downloading k8s.io/api v0.28.6
go: downloading github.com/emirpasic/gods v1.18.1
go: downloading github.com/aws/aws-sdk-go-v2/internal/endpoints/v2 v2.4.30
go: downloading github.com/acarl005/stripansi v0.0.0-20180116102854-5a71ef0e047d
go: downloading github.com/rivo/uniseg v0.4.7
go: downloading github.com/golang/glog v1.2.0
go: downloading github.com/lestrrat-go/httpcc v1.0.1
go: downloading github.com/ncw/directio v1.0.5
go: downloading github.com/coocood/rtutil v0.0.0-20190304133409-c84515f646f2
go: downloading github.com/coocood/bbloom v0.0.0-20190830030839-58deb6228d64
go: downloading github.com/klauspost/cpuid v1.3.1
go: downloading k8s.io/apimachinery v0.28.6
go: downloading gopkg.in/inf.v0 v0.9.1
go: downloading k8s.io/klog/v2 v2.120.1
go: downloading github.com/google/gofuzz v1.2.0
go: downloading k8s.io/utils v0.0.0-20230726121419-3b25d923346b
go: downloading sigs.k8s.io/structured-merge-diff/v4 v4.4.1
go: downloading sigs.k8s.io/json v0.0.0-20221116044647-bc3834ca7abd
go: downloading modernc.org/memory v1.7.2
go: downloading modernc.org/mathutil v1.6.0
go: downloading github.com/ardielle/ardielle-go v1.5.2
go: downloading github.com/jmespath/go-jmespath v0.4.0
go: downloading github.com/google/s2a-go v0.1.7
go: downloading github.com/googleapis/enterprise-certificate-proxy v0.3.2
go: downloading go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp v0.49.0
go: downloading github.com/felixge/httpsnoop v1.0.4
+ '[' -f ./bin/cdc_kafka_consumer ']'
+ make kafka_consumer
CGO_ENABLED=0 GO111MODULE=on go build  -trimpath  -ldflags '-X "github.com/pingcap/tiflow/pkg/version.ReleaseVersion=v8.2.0-alpha-88-g57a3c8e82" -X "github.com/pingcap/tiflow/pkg/version.BuildTS=2024-05-04 11:31:04" -X "github.com/pingcap/tiflow/pkg/version.GitHash=57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce" -X "github.com/pingcap/tiflow/pkg/version.GitBranch=HEAD" -X "github.com/pingcap/tiflow/pkg/version.GoVersion=go version go1.21.0 linux/amd64" -X "github.com/pingcap/tidb/pkg/parser/mysql.TiDBReleaseVersion=v8.2.0-alpha-88-g57a3c8e82"' -o bin/cdc_kafka_consumer ./cmd/kafka-consumer/main.go
+ '[' -f ./bin/cdc_storage_consumer ']'
+ make storage_consumer
CGO_ENABLED=0 GO111MODULE=on go build  -trimpath  -ldflags '-X "github.com/pingcap/tiflow/pkg/version.ReleaseVersion=v8.2.0-alpha-88-g57a3c8e82" -X "github.com/pingcap/tiflow/pkg/version.BuildTS=2024-05-04 11:31:11" -X "github.com/pingcap/tiflow/pkg/version.GitHash=57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce" -X "github.com/pingcap/tiflow/pkg/version.GitBranch=HEAD" -X "github.com/pingcap/tiflow/pkg/version.GoVersion=go version go1.21.0 linux/amd64" -X "github.com/pingcap/tidb/pkg/parser/mysql.TiDBReleaseVersion=v8.2.0-alpha-88-g57a3c8e82"' -o bin/cdc_storage_consumer ./cmd/storage-consumer/main.go
+ '[' -f ./bin/cdc.test ']'
+ make integration_test_build
cd tools/check && GO111MODULE=on go build -mod=mod -o ../bin/failpoint-ctl github.com/pingcap/failpoint/failpoint-ctl
go: downloading github.com/pingcap/failpoint v0.0.0-20210316064728-7acb0f0a3dfd
go: downloading github.com/sergi/go-diff v1.1.0
CGO_ENABLED=0 GO111MODULE=on go build  -trimpath  -ldflags '-X "github.com/pingcap/tiflow/pkg/version.ReleaseVersion=v8.2.0-alpha-88-g57a3c8e82" -X "github.com/pingcap/tiflow/pkg/version.BuildTS=2024-05-04 11:31:18" -X "github.com/pingcap/tiflow/pkg/version.GitHash=57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce" -X "github.com/pingcap/tiflow/pkg/version.GitBranch=HEAD" -X "github.com/pingcap/tiflow/pkg/version.GoVersion=go version go1.21.0 linux/amd64" -X "github.com/pingcap/tidb/pkg/parser/mysql.TiDBReleaseVersion=v8.2.0-alpha-88-g57a3c8e82"' -o bin/cdc_storage_consumer ./cmd/storage-consumer/main.go
CGO_ENABLED=0 GO111MODULE=on go build  -trimpath  -ldflags '-X "github.com/pingcap/tiflow/pkg/version.ReleaseVersion=v8.2.0-alpha-88-g57a3c8e82" -X "github.com/pingcap/tiflow/pkg/version.BuildTS=2024-05-04 11:31:18" -X "github.com/pingcap/tiflow/pkg/version.GitHash=57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce" -X "github.com/pingcap/tiflow/pkg/version.GitBranch=HEAD" -X "github.com/pingcap/tiflow/pkg/version.GoVersion=go version go1.21.0 linux/amd64" -X "github.com/pingcap/tidb/pkg/parser/mysql.TiDBReleaseVersion=v8.2.0-alpha-88-g57a3c8e82"' -o bin/cdc_kafka_consumer ./cmd/kafka-consumer/main.go
CGO_ENABLED=0 GO111MODULE=on go build  -trimpath  -ldflags '-X "github.com/pingcap/tiflow/pkg/version.ReleaseVersion=v8.2.0-alpha-88-g57a3c8e82" -X "github.com/pingcap/tiflow/pkg/version.BuildTS=2024-05-04 11:31:18" -X "github.com/pingcap/tiflow/pkg/version.GitHash=57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce" -X "github.com/pingcap/tiflow/pkg/version.GitBranch=HEAD" -X "github.com/pingcap/tiflow/pkg/version.GoVersion=go version go1.21.0 linux/amd64" -X "github.com/pingcap/tidb/pkg/parser/mysql.TiDBReleaseVersion=v8.2.0-alpha-88-g57a3c8e82"' -o bin/cdc_pulsar_consumer ./cmd/pulsar-consumer/main.go
CGO_ENABLED=0 GO111MODULE=on go build  -trimpath  -ldflags '-X "github.com/pingcap/tiflow/pkg/version.ReleaseVersion=v8.2.0-alpha-88-g57a3c8e82" -X "github.com/pingcap/tiflow/pkg/version.BuildTS=2024-05-04 11:31:18" -X "github.com/pingcap/tiflow/pkg/version.GitHash=57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce" -X "github.com/pingcap/tiflow/pkg/version.GitBranch=HEAD" -X "github.com/pingcap/tiflow/pkg/version.GoVersion=go version go1.21.0 linux/amd64" -X "github.com/pingcap/tidb/pkg/parser/mysql.TiDBReleaseVersion=v8.2.0-alpha-88-g57a3c8e82"' -o bin/oauth2-server ./cmd/oauth2-server/main.go
go: downloading github.com/go-oauth2/oauth2/v4 v4.5.2
go: downloading github.com/tidwall/buntdb v1.3.0
go: downloading github.com/tidwall/rtred v0.1.2
go: downloading github.com/tidwall/gjson v1.14.3
go: downloading github.com/tidwall/match v1.1.1
go: downloading github.com/tidwall/grect v0.1.4
go: downloading github.com/tidwall/tinyqueue v0.1.1
go: downloading github.com/tidwall/pretty v1.2.0
$(echo $(for p in $(go list ./... | grep -vE 'vendor|proto|tiflow/tests|integration|testing_utils|pb|pbmock|tiflow/bin'); do echo ${p#"github.com/pingcap/tiflow/"}|grep -v "github.com/pingcap/tiflow"; done) | xargs tools/bin/failpoint-ctl enable >/dev/null)
go: downloading github.com/PingCAP-QE/go-sqlsmith v0.0.0-20231213065948-336e064b488d
go: downloading github.com/chzyer/readline v1.5.1
go: downloading github.com/deepmap/oapi-codegen v1.9.0
go: downloading github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510
go: downloading github.com/getkin/kin-openapi v0.80.0
go: downloading github.com/gogo/gateway v1.1.0
go: downloading github.com/syndtr/goleveldb v1.0.1-0.20210305035536-64b5b1c73954
go: downloading go.uber.org/dig v1.13.0
go: downloading github.com/mattn/go-shellwords v1.0.12
go: downloading go.uber.org/ratelimit v0.2.0
go: downloading github.com/VividCortex/mysqlerr v1.0.0
go: downloading go.uber.org/goleak v1.3.0
go: downloading github.com/bradleyjkemp/grpc-tools v0.2.5
go: downloading github.com/integralist/go-findroot v0.0.0-20160518114804-ac90681525dc
go: downloading github.com/jmoiron/sqlx v1.3.3
go: downloading upper.io/db.v3 v3.7.1+incompatible
go: downloading github.com/DataDog/zstd v1.5.5
go: downloading github.com/shurcooL/httpgzip v0.0.0-20190720172056-320755c1c1b0
go: downloading github.com/andres-erbsen/clock v0.0.0-20160526145045-9e14626cd129
go: downloading github.com/ngaut/log v0.0.0-20210830112240-0124ec040aeb
go: downloading github.com/improbable-eng/grpc-web v0.12.0
go: downloading github.com/ghodss/yaml v1.0.0
go: downloading github.com/rs/cors v1.7.0
go: downloading github.com/desertbit/timer v0.0.0-20180107155436-c41aec40b27f
CGO_ENABLED=1 GO111MODULE=on go test -p 3 --race --tags=intest -ldflags '-X "github.com/pingcap/tiflow/pkg/version.ReleaseVersion=v8.2.0-alpha-88-g57a3c8e82" -X "github.com/pingcap/tiflow/pkg/version.BuildTS=2024-05-04 11:31:18" -X "github.com/pingcap/tiflow/pkg/version.GitHash=57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce" -X "github.com/pingcap/tiflow/pkg/version.GitBranch=HEAD" -X "github.com/pingcap/tiflow/pkg/version.GoVersion=go version go1.21.0 linux/amd64" -X "github.com/pingcap/tidb/pkg/parser/mysql.TiDBReleaseVersion=v8.2.0-alpha-88-g57a3c8e82"' -c -cover -covermode=atomic \
	-coverpkg=github.com/pingcap/tiflow/... \
	-o bin/cdc.test github.com/pingcap/tiflow/cmd/cdc \
|| { $(echo $(for p in $(go list ./... | grep -vE 'vendor|proto|tiflow/tests|integration|testing_utils|pb|pbmock|tiflow/bin'); do echo ${p#"github.com/pingcap/tiflow/"}|grep -v "github.com/pingcap/tiflow"; done) | xargs tools/bin/failpoint-ctl disable >/dev/null); exit 1; }
CGO_ENABLED=0 GO111MODULE=on go build  -trimpath  -ldflags '-X "github.com/pingcap/tiflow/pkg/version.ReleaseVersion=v8.2.0-alpha-88-g57a3c8e82" -X "github.com/pingcap/tiflow/pkg/version.BuildTS=2024-05-04 11:31:18" -X "github.com/pingcap/tiflow/pkg/version.GitHash=57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce" -X "github.com/pingcap/tiflow/pkg/version.GitBranch=HEAD" -X "github.com/pingcap/tiflow/pkg/version.GoVersion=go version go1.21.0 linux/amd64" -X "github.com/pingcap/tidb/pkg/parser/mysql.TiDBReleaseVersion=v8.2.0-alpha-88-g57a3c8e82"' -o bin/cdc ./cmd/cdc/main.go \
|| { $(echo $(for p in $(go list ./... | grep -vE 'vendor|proto|tiflow/tests|integration|testing_utils|pb|pbmock|tiflow/bin'); do echo ${p#"github.com/pingcap/tiflow/"}|grep -v "github.com/pingcap/tiflow"; done) | xargs tools/bin/failpoint-ctl disable >/dev/null); exit 1; }
$(echo $(for p in $(go list ./... | grep -vE 'vendor|proto|tiflow/tests|integration|testing_utils|pb|pbmock|tiflow/bin'); do echo ${p#"github.com/pingcap/tiflow/"}|grep -v "github.com/pingcap/tiflow"; done) | xargs tools/bin/failpoint-ctl disable >/dev/null)
+ ls -alh ./bin
total 1.2G
drwxr-sr-x  2 jenkins jenkins 4.0K May  4 19:36 .
drwxr-sr-x 19 jenkins jenkins 4.0K May  4 19:29 ..
-rwxr-xr-x  1 jenkins jenkins 220M May  4 19:36 cdc
-rwxr-xr-x  1 jenkins jenkins 359M May  4 19:35 cdc.test
-rwxr-xr-x  1 jenkins jenkins 183M May  4 19:31 cdc_kafka_consumer
-rwxr-xr-x  1 jenkins jenkins 183M May  4 19:31 cdc_pulsar_consumer
-rwxr-xr-x  1 jenkins jenkins 182M May  4 19:31 cdc_storage_consumer
-rwxr-xr-x  1 jenkins jenkins  12M May  4 19:31 oauth2-server
+ ./bin/cdc version
Release Version: v8.2.0-alpha-88-g57a3c8e82
Git Commit Hash: 57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce
Git Branch: HEAD
UTC Build Time: 2024-05-04 11:31:18
Go Version: go version go1.21.0 linux/amd64
Failpoint Build: true
[Pipeline] }
Cache saved successfully (binary/pingcap/tiflow/cdc-integration-test/rev-be15534-57a3c8e)
1191721984 bytes in 36.14 secs (32977407 bytes/sec)
[Pipeline] // cache
[Pipeline] cache
Cache not restored (no such key found)
[Pipeline] {
[Pipeline] sh
+ cp -r ../third_party_download/bin/bin ../third_party_download/bin/etc ../third_party_download/bin/etcdctl ../third_party_download/bin/go-ycsb ../third_party_download/bin/jq ../third_party_download/bin/lib ../third_party_download/bin/libc++.so.1 ../third_party_download/bin/libc++.so.1.0 ../third_party_download/bin/libc++abi.so.1 ../third_party_download/bin/libc++abi.so.1.0 ../third_party_download/bin/libgmssl.so ../third_party_download/bin/libgmssl.so.3 ../third_party_download/bin/libgmssl.so.3.0 ../third_party_download/bin/libtiflash_proxy.so ../third_party_download/bin/minio ../third_party_download/bin/pd-api-bench ../third_party_download/bin/pd-ctl ../third_party_download/bin/pd-heartbeat-bench ../third_party_download/bin/pd-recover ../third_party_download/bin/pd-server ../third_party_download/bin/pd-tso-bench ../third_party_download/bin/pd-ut ../third_party_download/bin/regions-dump ../third_party_download/bin/share ../third_party_download/bin/stores-dump ../third_party_download/bin/sync_diff_inspector ../third_party_download/bin/tidb-server ../third_party_download/bin/tiflash ../third_party_download/bin/tikv-server ../third_party_download/bin/xprog ./bin/
+ ls -alh ./bin
total 3.0G
drwxr-sr-x  6 jenkins jenkins  4.0K May  4 19:36 .
drwxr-sr-x 19 jenkins jenkins  4.0K May  4 19:29 ..
drwxr-sr-x  2 jenkins jenkins  4.0K May  4 19:36 bin
-rwxr-xr-x  1 jenkins jenkins  220M May  4 19:36 cdc
-rwxr-xr-x  1 jenkins jenkins  359M May  4 19:35 cdc.test
-rwxr-xr-x  1 jenkins jenkins  183M May  4 19:31 cdc_kafka_consumer
-rwxr-xr-x  1 jenkins jenkins  183M May  4 19:31 cdc_pulsar_consumer
-rwxr-xr-x  1 jenkins jenkins  182M May  4 19:31 cdc_storage_consumer
drwxr-sr-x  4 jenkins jenkins  4.0K May  4 19:36 etc
-rwxr-xr-x  1 jenkins jenkins   17M May  4 19:36 etcdctl
-rwxr-xr-x  1 jenkins jenkins   44M May  4 19:36 go-ycsb
-rwxr-xr-x  1 jenkins jenkins  3.8M May  4 19:36 jq
drwxr-sr-x  3 jenkins jenkins  4.0K May  4 19:36 lib
lrwxrwxrwx  1 jenkins jenkins    13 May  4 19:36 libc++.so.1 -> libc++.so.1.0
-rwxr-xr-x  1 jenkins jenkins 1016K May  4 19:36 libc++.so.1.0
lrwxrwxrwx  1 jenkins jenkins    16 May  4 19:36 libc++abi.so.1 -> libc++abi.so.1.0
-rwxr-xr-x  1 jenkins jenkins  358K May  4 19:36 libc++abi.so.1.0
lrwxrwxrwx  1 jenkins jenkins    13 May  4 19:36 libgmssl.so -> libgmssl.so.3
lrwxrwxrwx  1 jenkins jenkins    15 May  4 19:36 libgmssl.so.3 -> libgmssl.so.3.0
-rwxr-xr-x  1 jenkins jenkins  2.6M May  4 19:36 libgmssl.so.3.0
-rwxr-xr-x  1 jenkins jenkins  272M May  4 19:36 libtiflash_proxy.so
-rwxr-xr-x  1 jenkins jenkins   50M May  4 19:36 minio
-rwxr-xr-x  1 jenkins jenkins   12M May  4 19:31 oauth2-server
-rwxr-xr-x  1 jenkins jenkins   37M May  4 19:36 pd-api-bench
-rwxr-xr-x  1 jenkins jenkins   44M May  4 19:36 pd-ctl
-rwxr-xr-x  1 jenkins jenkins   36M May  4 19:36 pd-heartbeat-bench
-rwxr-xr-x  1 jenkins jenkins   32M May  4 19:36 pd-recover
-rwxr-xr-x  1 jenkins jenkins  106M May  4 19:36 pd-server
-rwxr-xr-x  1 jenkins jenkins   26M May  4 19:36 pd-tso-bench
-rwxr-xr-x  1 jenkins jenkins  3.0M May  4 19:36 pd-ut
-rwxr-xr-x  1 jenkins jenkins   32M May  4 19:36 regions-dump
drwxr-sr-x  4 jenkins jenkins  4.0K May  4 19:36 share
-rwxr-xr-x  1 jenkins jenkins   32M May  4 19:36 stores-dump
-rwxr-xr-x  1 jenkins jenkins  192M May  4 19:36 sync_diff_inspector
-rwxr-xr-x  1 jenkins jenkins  208M May  4 19:36 tidb-server
-rwxr-xr-x  1 jenkins jenkins  380M May  4 19:36 tiflash
-rwxr-xr-x  1 jenkins jenkins  418M May  4 19:36 tikv-server
-rwxr-xr-x  1 jenkins jenkins  2.0M May  4 19:36 xprog
[Pipeline] }
Cache saved successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 60.50 secs (61605254 bytes/sec)
[Pipeline] // cache
[Pipeline] }
[Pipeline] // dir
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Tests)
[Pipeline] parallel
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G00')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G01')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G02')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G03')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G04')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G05')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G06')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G07')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G08')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G09')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G10')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G11')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G12')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G13')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G14')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G15')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G16')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G17')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G00')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G01')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G02')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G03')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G04')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G05')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G06')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G07')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G08')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G09')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G10')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G11')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G12')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G13')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G14')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G15')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G16')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G17')
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
[Pipeline] readTrusted
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-ttdhm-2w16j
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "cbbf42b5100a022b6fc749dafa541767ce105c00"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-ttdhm"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
[Pipeline] {
[Pipeline] checkout
The recommended git tool is: git
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@d2f8941; decorates RemoteLauncher[hudson.remoting.Channel@71ed5a65:JNLP4-connect connection from 10.233.69.225/10.233.69.225:50250] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
Cloning repository https://github.com/PingCAP-QE/ci.git
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf
[Pipeline] node
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-8dbxr-5vmhm
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "d1315c497a493e5dbd0d935413cd558256ba1ab2"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-8dbxr"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
[Pipeline] {
[Pipeline] checkout
The recommended git tool is: git
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-pfc68-v3c7z
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "d1f73cb7b02a4bf33dd4010c081bade4040d269f"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-pfc68"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk
Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
[Pipeline] {
[Pipeline] checkout
The recommended git tool is: git
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@7ad7051c; decorates RemoteLauncher[hudson.remoting.Channel@7ed6d002:JNLP4-connect connection from 10.233.66.242/10.233.66.242:55750] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@378eda34; decorates RemoteLauncher[hudson.remoting.Channel@8bf1d66:JNLP4-connect connection from 10.233.70.88/10.233.70.88:34074] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-bft7v-gt97x
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "b050f074d49c43d30c20031e72d413dcae6b5ad0"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-bft7v"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
[Pipeline] {
[Pipeline] checkout
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
The recommended git tool is: git
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7 is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-gc0bt-l557l
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "fae64df29f14cf1ffac2788a7e5a8b8e8630305e"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-gc0bt"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7 in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
[Pipeline] {
[Pipeline] checkout
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
The recommended git tool is: git
[Pipeline] withEnv
[Pipeline] {
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
[Pipeline] container
[Pipeline] {
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] cache
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@56d48f6c; decorates RemoteLauncher[hudson.remoting.Channel@2239759:JNLP4-connect connection from 10.233.107.229/10.233.107.229:52448] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@7ad4cac3; decorates RemoteLauncher[hudson.remoting.Channel@32af4149:JNLP4-connect connection from 10.233.67.187/10.233.67.187:39562] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44 is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-xx2km-855hr
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "4e333fde59586b78f3d7bcbf70a2f448879e1129"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-xx2km"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44 in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-m43dx-nzz01
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "86dc05ac4b13fca41986f901911d5a36b5a7340b"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-m43dx"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-ftwvg-sd287
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "98b441031e8fa5f92ab4262be5551f25816a3755"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-ftwvg"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-t0ptc-lsf4m
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "8ca8a65103d46643ec7f5a963d0ed8405f15f2f9"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-t0ptc"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Still waiting to schedule task
โ€˜pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpdโ€™ is offline
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-gzt73-f2lwf
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "4eb63d7ff8425ec55f54034682db36265664ae97"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-gzt73"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-6tpkp-blzgt
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "84ef9a3d10dbf660de79e6a06ba9a739980b6319"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-6tpkp"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-321md-3zzc5
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "1f7c83d9553a137e9a9e80f411e98278dda144b4"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-321md"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 34.45 secs (108184154 bytes/sec)
[Pipeline] {
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] {
[Pipeline] node
[Pipeline] node
[Pipeline] node
[Pipeline] node
[Pipeline] node
[Pipeline] node
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] sh
[Pipeline] checkout
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
The recommended git tool is: git
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-tgsj5-209qp
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-38244-hqb2r
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d59wt-f77sf
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-s7s3l-7txfz
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6gzn1-2sfnd
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d1pdj-fsd6q
[Pipeline] checkout
[Pipeline] checkout
The recommended git tool is: git
[Pipeline] checkout
The recommended git tool is: git
[Pipeline] checkout
The recommended git tool is: git
[Pipeline] checkout
The recommended git tool is: git
[Pipeline] checkout
The recommended git tool is: git
The recommended git tool is: git
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] }
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@a883887; decorates RemoteLauncher[hudson.remoting.Channel@46aa90fe:JNLP4-connect connection from 10.233.71.22/10.233.71.22:48476] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] // timeout
[Pipeline] }
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] // container
[Pipeline] sh
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@582711df; decorates RemoteLauncher[hudson.remoting.Channel@1c770c09:JNLP4-connect connection from 10.233.105.113/10.233.105.113:51278] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@47fe0b6d; decorates RemoteLauncher[hudson.remoting.Channel@8e63777:JNLP4-connect connection from 10.233.84.68/10.233.84.68:39794] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@4c1a5cae; decorates RemoteLauncher[hudson.remoting.Channel@49abb7b5:JNLP4-connect connection from 10.233.108.66/10.233.108.66:49738] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@58deba57; decorates RemoteLauncher[hudson.remoting.Channel@2ae72c59:JNLP4-connect connection from 10.233.88.214/10.233.88.214:43396] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@16a008c3; decorates RemoteLauncher[hudson.remoting.Channel@718df3fc:JNLP4-connect connection from 10.233.127.138/10.233.127.138:40408] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@42b610dd; decorates RemoteLauncher[hudson.remoting.Channel@191ddc87:JNLP4-connect connection from 10.233.72.137/10.233.72.137:45220] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
[Pipeline] cache
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G01
Run cases: open_protocol_handle_key_only
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=e456a74a-b074-4637-ad35-1d8f2859254c
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G01
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-ttdhm
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-ttdhm pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/open_protocol_handle_key_only/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
Avoid second fetch
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-38244-hqb2r is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-38244-dljf2
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "693f732973d28110a707d42622a4bce7653eb561"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-38244"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-38244-hqb2r"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-38244-hqb2r"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-38244-hqb2r"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-38244-hqb2r in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
start tidb cluster in /tmp/tidb_cdc_test/open_protocol_handle_key_only
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d1pdj-fsd6q is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-d1pdj-wpvxm
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "62d84c0acba04ce3afa9a855302a2f8576f98200"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-d1pdj"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d1pdj-fsd6q"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d1pdj-fsd6q"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d1pdj-fsd6q"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d1pdj-fsd6q in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-tgsj5-209qp is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-tgsj5-xmnfz
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "4901a9b4e334bc90db484e3682008d4b7ad5b0ba"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-tgsj5"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-tgsj5-209qp"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-tgsj5-209qp"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-tgsj5-209qp"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-tgsj5-209qp in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d59wt-f77sf is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-d59wt-j3d8q
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "10e96fa43c4c2bdfc39e758cf4326bb21817c310"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-d59wt"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d59wt-f77sf"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d59wt-f77sf"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d59wt-f77sf"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d59wt-f77sf in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-s7s3l-7txfz is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-s7s3l-8s6sx
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "15dcdbc2df259d81d5fd299f4bbbe6549b20ebcd"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-s7s3l"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-s7s3l-7txfz"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-s7s3l-7txfz"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-s7s3l-7txfz"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Avoid second fetch
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-s7s3l-7txfz in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6gzn1-2sfnd is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1845-6gzn1-nq8l9
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "5c23c60580e13ce5cbd8c433e644b3b0ddb09e67"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1845-6gzn1"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6gzn1-2sfnd"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6gzn1-2sfnd"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6gzn1-2sfnd"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6gzn1-2sfnd in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 8.56 secs (435510494 bytes/sec)
[Pipeline] {
[Pipeline] cache
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9a15040014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g, pid:1355, start at 2024-05-04 19:39:03.408416439 +0800 CST m=+5.254104697	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:41:03.415 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:39:03.411 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:29:03.411 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9a15040014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g, pid:1355, start at 2024-05-04 19:39:03.408416439 +0800 CST m=+5.254104697	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:41:03.415 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:39:03.411 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:29:03.411 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9a15a80017	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ttdhm-czj8g, pid:1433, start at 2024-05-04 19:39:03.450990441 +0800 CST m=+5.245499049	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:41:03.459 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:39:03.452 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:29:03.452 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/log/error.log
arg matches is ArgMatches { args: {"data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/db/proxy"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/log/proxy.log"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash-proxy.toml"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.open_protocol_handle_key_only.cli.2824.out cli tso query --pd=http://127.0.0.1:2379
+ set +x
+ tso='449530494664310787
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530494664310787 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
[Sat May  4 19:39:10 CST 2024] <<<<<< START cdc server in open_protocol_handle_key_only case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ GO_FAILPOINTS=
+ (( i = 0 ))
+ (( i <= 50 ))
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.open_protocol_handle_key_only.28602862.out server --log-file /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data --cluster-id default
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 9.74 secs (382819428 bytes/sec)
[Pipeline] {
[Pipeline] cache
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:39:13 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/5c145f57-5f64-47b4-b31b-1b5a1d42d606
	{"id":"5c145f57-5f64-47b4-b31b-1b5a1d42d606","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822750}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43665b0cd3
	5c145f57-5f64-47b4-b31b-1b5a1d42d606

/tidb/cdc/default/default/upstream/7365107569143670156
	{"id":7365107569143670156,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/5c145f57-5f64-47b4-b31b-1b5a1d42d606
	{"id":"5c145f57-5f64-47b4-b31b-1b5a1d42d606","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822750}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43665b0cd3
	5c145f57-5f64-47b4-b31b-1b5a1d42d606

/tidb/cdc/default/default/upstream/7365107569143670156
	{"id":7365107569143670156,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/5c145f57-5f64-47b4-b31b-1b5a1d42d606
	{"id":"5c145f57-5f64-47b4-b31b-1b5a1d42d606","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822750}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43665b0cd3
	5c145f57-5f64-47b4-b31b-1b5a1d42d606

/tidb/cdc/default/default/upstream/7365107569143670156
	{"id":7365107569143670156,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.open_protocol_handle_key_only.cli.2918.out cli changefeed create --start-ts=449530494664310787 '--sink-uri=kafka://127.0.0.1:9092/open-protocol-handle-key-only?protocol=open-protocol&max-message-bytes=800&kafka-version=2.4.1' --config=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/open_protocol_handle_key_only/conf/changefeed.toml
Create changefeed successfully!
ID: 036518b8-be84-4454-ab8c-3a74b13f37b9
Info: {"upstream_id":7365107569143670156,"namespace":"default","id":"036518b8-be84-4454-ab8c-3a74b13f37b9","sink_uri":"kafka://127.0.0.1:9092/open-protocol-handle-key-only?protocol=open-protocol\u0026max-message-bytes=800\u0026kafka-version=2.4.1","create_time":"2024-05-04T19:39:13.894727847+08:00","start_ts":449530494664310787,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"kafka_config":{"large_message_handle":{"large_message_handle_option":"handle-key-only","large_message_handle_compression":"lz4","claim_check_storage_uri":""}},"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530494664310787,"checkpoint_ts":449530494664310787,"checkpoint_time":"2024-05-04 19:39:08.811"}
PASS
coverage: 2.5% of statements in github.com/pingcap/tiflow/...
+ set +x
table test.finish_mark not exists for 1-th check, retry later
table test.finish_mark not exists for 2-th check, retry later
table test.finish_mark exists
check diff failed 1-th time, retry later
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 7.05 secs (528525496 bytes/sec)
[Pipeline] {
[Pipeline] cache
check diff failed 2-th time, retry later
check diff failed 3-th time, retry later
check diff failed 4-th time, retry later
check diff failed 5-th time, retry later
check diff failed 6-th time, retry later
check diff failed 7-th time, retry later
check diff failed 8-th time, retry later
check diff failed 9-th time, retry later
check diff failed 10-th time, retry later
check diff failed at last
A total of 2 tables need to be compared


Comparing the table structure of ``test`.`t`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/1
Comparing the table structure of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/2
Comparing the table data of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [==============================>------------------------------] 50% 0/1
Comparing the table data of ``test`.`t`` ... failure
_____________________________________________________________________________
Progress [============================================================>] 100% 0/0
Progress [============================================================>] 100% 0/0
The data of `test`.`t` is not equal

The rest of tables are all equal.

A total of 2 tables have been compared, 1 tables finished, 1 tables failed, 0 tables skipped.
The patch file has been generated in 
	'/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/fix-on-tidb0/'
You can view the comparision details through '/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log'
A total of 2 tables need to be compared


Comparing the table structure of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/1
Comparing the table structure of ``test`.`t`` ... equivalent
Comparing the table data of ``test`.`t`` ...
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/2
Comparing the table data of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [==============================>------------------------------] 50% 0/1
Comparing the table data of ``test`.`t`` ... failure
_____________________________________________________________________________
Progress [============================================================>] 100% 0/0
Progress [============================================================>] 100% 0/0
The data of `test`.`t` is not equal

The rest of tables are all equal.

A total of 2 tables have been compared, 1 tables finished, 1 tables failed, 0 tables skipped.
The patch file has been generated in 
	'/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/fix-on-tidb0/'
You can view the comparision details through '/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log'
A total of 2 tables need to be compared


Comparing the table structure of ``test`.`t`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/1
Comparing the table structure of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/2
Comparing the table data of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [==============================>------------------------------] 50% 0/1
Comparing the table data of ``test`.`t`` ... failure
_____________________________________________________________________________
Progress [============================================================>] 100% 0/0
Progress [============================================================>] 100% 0/0
The data of `test`.`t` is not equal

The rest of tables are all equal.

A total of 2 tables have been compared, 1 tables finished, 1 tables failed, 0 tables skipped.
The patch file has been generated in 
	'/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/fix-on-tidb0/'
You can view the comparision details through '/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log'
A total of 2 tables need to be compared


Comparing the table structure of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/1
Comparing the table structure of ``test`.`t`` ... equivalent
Comparing the table data of ``test`.`t`` ...
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/2
Comparing the table data of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [==============================>------------------------------] 50% 0/1
Comparing the table data of ``test`.`t`` ... failure
_____________________________________________________________________________
Progress [============================================================>] 100% 0/0
Progress [============================================================>] 100% 0/0
The data of `test`.`t` is not equal

The rest of tables are all equal.

A total of 2 tables have been compared, 1 tables finished, 1 tables failed, 0 tables skipped.
The patch file has been generated in 
	'/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/fix-on-tidb0/'
You can view the comparision details through '/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log'
A total of 2 tables need to be compared


Comparing the table structure of ``test`.`t`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/1
Comparing the table structure of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/2
Comparing the table data of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [==============================>------------------------------] 50% 0/1
Comparing the table data of ``test`.`t`` ... failure
_____________________________________________________________________________
Progress [============================================================>] 100% 0/0
Progress [============================================================>] 100% 0/0
The data of `test`.`t` is not equal

The rest of tables are all equal.

A total of 2 tables have been compared, 1 tables finished, 1 tables failed, 0 tables skipped.
The patch file has been generated in 
	'/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/fix-on-tidb0/'
You can view the comparision details through '/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log'
A total of 2 tables need to be compared


Comparing the table structure of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/1
Comparing the table structure of ``test`.`t`` ... equivalent
Comparing the table data of ``test`.`t`` ...
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/2
Comparing the table data of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [==============================>------------------------------] 50% 0/1
Comparing the table data of ``test`.`t`` ... failure
_____________________________________________________________________________
Progress [============================================================>] 100% 0/0
Progress [============================================================>] 100% 0/0
The data of `test`.`t` is not equal

The rest of tables are all equal.

A total of 2 tables have been compared, 1 tables finished, 1 tables failed, 0 tables skipped.
The patch file has been generated in 
	'/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/fix-on-tidb0/'
You can view the comparision details through '/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log'
A total of 2 tables need to be compared


Comparing the table structure of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/1
Comparing the table structure of ``test`.`t`` ... equivalent
Comparing the table data of ``test`.`t`` ...
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/2
Comparing the table data of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [==============================>------------------------------] 50% 0/1
Comparing the table data of ``test`.`t`` ... failure
_____________________________________________________________________________
Progress [============================================================>] 100% 0/0
Progress [============================================================>] 100% 0/0
The data of `test`.`t` is not equal

The rest of tables are all equal.

A total of 2 tables have been compared, 1 tables finished, 1 tables failed, 0 tables skipped.
The patch file has been generated in 
	'/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/fix-on-tidb0/'
You can view the comparision details through '/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log'
A total of 2 tables need to be compared


Comparing the table structure of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/1
Comparing the table structure of ``test`.`t`` ... equivalent
Comparing the table data of ``test`.`t`` ...
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/2
Comparing the table data of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [==============================>------------------------------] 50% 0/1
Comparing the table data of ``test`.`t`` ... failure
_____________________________________________________________________________
Progress [============================================================>] 100% 0/0
Progress [============================================================>] 100% 0/0
The data of `test`.`t` is not equal

The rest of tables are all equal.

A total of 2 tables have been compared, 1 tables finished, 1 tables failed, 0 tables skipped.
The patch file has been generated in 
	'/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/fix-on-tidb0/'
You can view the comparision details through '/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log'
A total of 2 tables need to be compared


Comparing the table structure of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/1
Comparing the table structure of ``test`.`t`` ... equivalent
Comparing the table data of ``test`.`t`` ...
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/2
Comparing the table data of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [==============================>------------------------------] 50% 0/1
Comparing the table data of ``test`.`t`` ... failure
_____________________________________________________________________________
Progress [============================================================>] 100% 0/0
Progress [============================================================>] 100% 0/0
The data of `test`.`t` is not equal

The rest of tables are all equal.

A total of 2 tables have been compared, 1 tables finished, 1 tables failed, 0 tables skipped.
The patch file has been generated in 
	'/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/fix-on-tidb0/'
You can view the comparision details through '/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log'
A total of 2 tables need to be compared


Comparing the table structure of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/1
Comparing the table structure of ``test`.`t`` ... equivalent
Comparing the table data of ``test`.`t`` ...
Comparing the table data of ``test`.`finish_mark`` ...
_____________________________________________________________________________
Progress [>------------------------------------------------------------] 0% 0/2
Comparing the table data of ``test`.`finish_mark`` ... equivalent
Comparing the table data of ``test`.`t`` ...
_____________________________________________________________________________
Progress [==============================>------------------------------] 50% 0/1
Comparing the table data of ``test`.`t`` ... failure
_____________________________________________________________________________
Progress [============================================================>] 100% 0/0
Progress [============================================================>] 100% 0/0
The data of `test`.`t` is not equal

The rest of tables are all equal.

A total of 2 tables have been compared, 1 tables finished, 1 tables failed, 0 tables skipped.
The patch file has been generated in 
	'/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/fix-on-tidb0/'
You can view the comparision details through '/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log'

cat: /tmp/tidb_cdc_test/open_protocol_handle_key_only/sync_diff/output/sync_diff.log: No such file or directory

Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 30.68 secs (121486971 bytes/sec)
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] sh
[Pipeline] sh
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
[Pipeline] sh
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
[Pipeline] sh
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/open_protocol_handle_key_only/run.sh: line 1:  2948 Killed                  cdc_kafka_consumer --upstream-uri $SINK_URI --downstream-uri="mysql://root@127.0.0.1:3306/?safe-mode=true&batch-dml-enable=false&multi-stmt-enable=false" --upstream-tidb-dsn="root@tcp(${UP_TIDB_HOST}:${UP_TIDB_PORT})/?" --config="$CUR/conf/changefeed.toml" 2>&1
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] checkout
[Pipeline] }
The recommended git tool is: git
[Pipeline] checkout
[Pipeline] checkout
The recommended git tool is: git
[Pipeline] checkout
The recommended git tool is: git
[Pipeline] checkout
The recommended git tool is: git
[Pipeline] checkout
The recommended git tool is: git
[Pipeline] stage
[Pipeline] { (Test)
The recommended git tool is: git
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // cache
[Pipeline] }
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // container
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@2e61af95; decorates RemoteLauncher[hudson.remoting.Channel@261ec171:JNLP4-connect connection from 10.233.108.89/10.233.108.89:50652] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
[Pipeline] // container
Cloning the remote Git repository
Using shallow clone with depth 1
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@36869435; decorates RemoteLauncher[hudson.remoting.Channel@613d9251:JNLP4-connect connection from 10.233.123.199/10.233.123.199:56374] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
[Pipeline] // container
[Pipeline] sh
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@5b5c8fc6; decorates RemoteLauncher[hudson.remoting.Channel@37dd1e4e:JNLP4-connect connection from 10.233.107.29/10.233.107.29:40710] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@2d70f4fb; decorates RemoteLauncher[hudson.remoting.Channel@10e7027:JNLP4-connect connection from 10.233.66.207/10.233.66.207:47618] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@1d08766e; decorates RemoteLauncher[hudson.remoting.Channel@fd3291b:JNLP4-connect connection from 10.233.73.69/10.233.73.69:59190] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
[Pipeline] sh
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
No credentials specified
Warning: JENKINS-30600: special launcher org.csanchez.jenkins.plugins.kubernetes.pipeline.ContainerExecDecorator$1@6b23595f; decorates RemoteLauncher[hudson.remoting.Channel@6982b837:JNLP4-connect connection from 10.233.71.160/10.233.71.160:51882] will be ignored (a typical symptom is the Git executable not being run inside a designated container)
Cloning the remote Git repository
Using shallow clone with depth 1
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G03
Run cases: row_format drop_many_tables processor_stop_delay partition_table
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=560a0a75-a82c-4306-bdac-eab219add4a2
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G03
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-8dbxr
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-8dbxr pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/row_format/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
[Pipeline] sh
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G04
Run cases: foreign_key ddl_puller_lag ddl_only_block_related_table changefeed_auto_stop
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=afad64ae-ba23-4a7b-8585-a6ccdad6e701
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G04
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-pfc68
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv pingcap_tiflow_pull_cdc_integration_kafka_test_1845-pfc68
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/foreign_key/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
[Pipeline] // dir
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
Post stage
[Pipeline] // container
[Pipeline] sh
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G05
Run cases: charset_gbk ddl_manager multi_source
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=6aa2aaf2-5242-474c-85fd-88e58d170ccd
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT=tcp://10.233.0.1:443
KUBERNETES_PORT_443_TCP_PORT=443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G05
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-bft7v
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-bft7v pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/charset_gbk/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] sh
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G00
Run cases: bdr_mode capture_suicide_while_balance_table syncpoint hang_sink_suicide server_config_compatibility changefeed_dup_error_restart kafka_big_messages kafka_compression kafka_messages kafka_sink_error_resume mq_sink_lost_callback mq_sink_dispatcher kafka_column_selector kafka_column_selector_avro debezium lossy_ddl storage_csv_update
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=4c4e0bb9-70c4-4e24-8987-ed61f1e93e5f
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G00
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-gc0bt
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-gc0bt pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/bdr_mode/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:40:01 CST 2024] <<<<<< run test case bdr_mode success! >>>>>>
[Pipeline] cache
+ ls /tmp/tidb_cdc_test/
cov.open_protocol_handle_key_only.cli.2824.out
cov.open_protocol_handle_key_only.cli.2918.out
open_protocol_handle_key_only
sql_res.open_protocol_handle_key_only.txt
++ find /tmp/tidb_cdc_test/ -type f -name '*.log'
+ tar -cvzf log-G01.tar.gz /tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv2.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv1.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/log/error.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/log/proxy.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/log/server.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/db/proxy/db/000005.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv3.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0005/000002.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0001/000002.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0006/000002.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0000/000002.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0002/000002.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0004/000002.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0003/000002.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0007/000002.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/stdout.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/down_pd/hot-region/000001.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/down_pd/region-meta/000001.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv3/db/000005.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tidb_other.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv_down.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv2/db/000005.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv1/db/000005.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/pd1/hot-region/000001.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/pd1/region-meta/000001.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tidb-slow.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_kafka_consumer.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/pd1.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv_down/db/000005.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tidb.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/down_pd.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/sync_diff_inspector.log /tmp/tidb_cdc_test/open_protocol_handle_key_only/tidb_down.log
tar: Removing leading `/' from member names
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv2.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv1.log
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/log/error.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/log/proxy.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/log/server.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tiflash/db/proxy/db/000005.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv3.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0005/000002.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0001/000002.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0006/000002.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0000/000002.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0002/000002.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0004/000002.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0003/000002.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_data/tmp/sorter/0007/000002.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/stdout.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/down_pd/hot-region/000001.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/down_pd/region-meta/000001.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv3/db/000005.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tidb_other.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv_down.log
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
/tmp/tidb_cdc_test/open_protocol_handle_key_only/output/sync_diff.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv2/db/000005.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv1/db/000005.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/pd1/hot-region/000001.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/pd1/region-meta/000001.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tidb-slow.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/cdc_kafka_consumer.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/pd1.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tikv_down/db/000005.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tidb.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/down_pd.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/sync_diff_inspector.log
/tmp/tidb_cdc_test/open_protocol_handle_key_only/tidb_down.log
+ ls -alh log-G01.tar.gz
-rw-r--r--. 1 jenkins jenkins 3.2M May  4 19:40 log-G01.tar.gz
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
start tidb cluster in /tmp/tidb_cdc_test/row_format
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
start tidb cluster in /tmp/tidb_cdc_test/foreign_key
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/capture_suicide_while_balance_table/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:40:04 CST 2024] <<<<<< run test case capture_suicide_while_balance_table success! >>>>>>
start tidb cluster in /tmp/tidb_cdc_test/charset_gbk
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Upstream TiDB...
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/syncpoint/run.sh using Sink-Type: kafka... <<=================
kafka downstream isn't support syncpoint record
[Sat May  4 19:40:07 CST 2024] <<<<<< run test case syncpoint success! >>>>>>
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/hang_sink_suicide/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:40:10 CST 2024] <<<<<< run test case hang_sink_suicide success! >>>>>>
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9e55e40002	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs, pid:1286, start at 2024-05-04 19:40:13.049759906 +0800 CST m=+5.111474394	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:13.055 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:13.049 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:13.049 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9e55e40002	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs, pid:1286, start at 2024-05-04 19:40:13.049759906 +0800 CST m=+5.111474394	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:13.055 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:13.049 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:13.049 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9e57680014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs, pid:1374, start at 2024-05-04 19:40:13.175934989 +0800 CST m=+5.168488724	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:13.182 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:13.146 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:13.146 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/charset_gbk/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/charset_gbk/tiflash/log/error.log
arg matches is ArgMatches { args: {"data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/charset_gbk/tiflash/db/proxy"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/charset_gbk/tiflash/log/proxy.log"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/charset_gbk/tiflash-proxy.toml"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/server_config_compatibility/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:40:13 CST 2024] <<<<<< run test case server_config_compatibility success! >>>>>>
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9e6b280013	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv, pid:1293, start at 2024-05-04 19:40:14.430986925 +0800 CST m=+5.078554740	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:14.438 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:14.410 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:14.410 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9e6b280013	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv, pid:1293, start at 2024-05-04 19:40:14.430986925 +0800 CST m=+5.078554740	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:14.438 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:14.410 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:14.410 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9e6ca0000b	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv, pid:1367, start at 2024-05-04 19:40:14.514132617 +0800 CST m=+5.111323367	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:14.523 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:14.504 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:14.504 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/foreign_key/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/foreign_key/tiflash/log/error.log
arg matches is ArgMatches { args: {"addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/foreign_key/tiflash/db/proxy"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/foreign_key/tiflash/log/proxy.log"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/foreign_key/tiflash-proxy.toml"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9e6950000b	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg, pid:1283, start at 2024-05-04 19:40:14.300624677 +0800 CST m=+5.172745920	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:14.309 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:14.292 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:14.292 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
[Sat May  4 19:40:16 CST 2024] <<<<<< START cdc server in charset_gbk case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ GO_FAILPOINTS=
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.charset_gbk.28412843.out server --log-file /tmp/tidb_cdc_test/charset_gbk/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/charset_gbk/cdc_data --cluster-id default
+ [[ no != \n\o ]]
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/changefeed_dup_error_restart/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:40:17 CST 2024] <<<<<< run test case changefeed_dup_error_restart success! >>>>>>
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9e6950000b	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg, pid:1283, start at 2024-05-04 19:40:14.300624677 +0800 CST m=+5.172745920	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:14.309 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:14.292 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:14.292 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9e688c0015	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-8dbxr-05ltg, pid:1375, start at 2024-05-04 19:40:14.278999953 +0800 CST m=+5.098594531	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:14.285 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:14.243 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:14.243 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/row_format/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/row_format/tiflash/log/error.log
arg matches is ArgMatches { args: {"pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/row_format/tiflash-proxy.toml"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/row_format/tiflash/log/proxy.log"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/row_format/tiflash/db/proxy"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.foreign_key.cli.2797.out cli tso query --pd=http://127.0.0.1:2379
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.row_format.cli.2837.out cli tso query --pd=http://127.0.0.1:2379
+ set +x
+ tso='449530512791306241
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530512791306241 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
[Sat May  4 19:40:19 CST 2024] <<<<<< START cdc server in foreign_key case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ GO_FAILPOINTS=
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.foreign_key.28342836.out server --log-file /tmp/tidb_cdc_test/foreign_key/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/foreign_key/cdc_data --cluster-id default
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:40:19 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/b7e55714-d149-4419-a342-f46f73ff9949
	{"id":"b7e55714-d149-4419-a342-f46f73ff9949","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822816}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43677378c3
	b7e55714-d149-4419-a342-f46f73ff9949

/tidb/cdc/default/default/upstream/7365107864998565802
	{"id":7365107864998565802,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/b7e55714-d149-4419-a342-f46f73ff9949
	{"id":"b7e55714-d149-4419-a342-f46f73ff9949","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822816}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43677378c3
	b7e55714-d149-4419-a342-f46f73ff9949

/tidb/cdc/default/default/upstream/7365107864998565802
	{"id":7365107864998565802,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/b7e55714-d149-4419-a342-f46f73ff9949
	{"id":"b7e55714-d149-4419-a342-f46f73ff9949","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822816}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43677378c3
	b7e55714-d149-4419-a342-f46f73ff9949

/tidb/cdc/default/default/upstream/7365107864998565802
	{"id":7365107864998565802,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
Create changefeed successfully!
ID: f43fc8f1-e40d-41ad-a165-9f141b374092
Info: {"upstream_id":7365107864998565802,"namespace":"default","id":"f43fc8f1-e40d-41ad-a165-9f141b374092","sink_uri":"mysql://normal:xxxxx@127.0.0.1:3306/","create_time":"2024-05-04T19:40:19.659052875+08:00","start_ts":449530512368730113,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530512368730113,"checkpoint_ts":449530512368730113,"checkpoint_time":"2024-05-04 19:40:16.348"}
[Sat May  4 19:40:19 CST 2024] <<<<<< START kafka consumer in charset_gbk case >>>>>>
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/kafka_big_messages/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
+ set +x
+ tso='449530513271554049
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530513271554049 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
[Sat May  4 19:40:21 CST 2024] <<<<<< START cdc server in row_format case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ GO_FAILPOINTS=
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.row_format.28752877.out server --log-file /tmp/tidb_cdc_test/row_format/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/row_format/cdc_data --cluster-id default
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:40:22 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/c54d8455-4f26-4dd0-8543-37cca6002095
	{"id":"c54d8455-4f26-4dd0-8543-37cca6002095","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822819}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4367712bcd
	c54d8455-4f26-4dd0-8543-37cca6002095

/tidb/cdc/default/default/upstream/7365107873093321768
	{"id":7365107873093321768,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/c54d8455-4f26-4dd0-8543-37cca6002095
	{"id":"c54d8455-4f26-4dd0-8543-37cca6002095","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822819}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4367712bcd
	c54d8455-4f26-4dd0-8543-37cca6002095

/tidb/cdc/default/default/upstream/7365107873093321768
	{"id":7365107873093321768,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/c54d8455-4f26-4dd0-8543-37cca6002095
	{"id":"c54d8455-4f26-4dd0-8543-37cca6002095","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822819}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4367712bcd
	c54d8455-4f26-4dd0-8543-37cca6002095

/tidb/cdc/default/default/upstream/7365107873093321768
	{"id":7365107873093321768,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.foreign_key.cli.2894.out cli changefeed create --start-ts=449530512791306241 '--sink-uri=kafka://127.0.0.1:9092/ticdc-foreign-key-test-20387?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760'
Create changefeed successfully!
ID: 35c3dbdc-5ce7-4520-9e45-3da97555aa26
Info: {"upstream_id":7365107873093321768,"namespace":"default","id":"35c3dbdc-5ce7-4520-9e45-3da97555aa26","sink_uri":"kafka://127.0.0.1:9092/ticdc-foreign-key-test-20387?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:40:22.984489658+08:00","start_ts":449530512791306241,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530512791306241,"checkpoint_ts":449530512791306241,"checkpoint_time":"2024-05-04 19:40:17.960"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
start tidb cluster in /tmp/tidb_cdc_test/kafka_big_messages
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
+ set +x
[Sat May  4 19:40:24 CST 2024] <<<<<< START kafka consumer in foreign_key case >>>>>>
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:40:24 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1e4981db-db53-4c47-be46-2f85b39c34a2
	{"id":"1e4981db-db53-4c47-be46-2f85b39c34a2","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822821}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43677052d4
	1e4981db-db53-4c47-be46-2f85b39c34a2

/tidb/cdc/default/default/upstream/7365107873580712052
	{"id":7365107873580712052,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1e4981db-db53-4c47-be46-2f85b39c34a2
	{"id":"1e4981db-db53-4c47-be46-2f85b39c34a2","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822821}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43677052d4
	1e4981db-db53-4c47-be46-2f85b39c34a2

/tidb/cdc/default/default/upstream/7365107873580712052
	{"id":7365107873580712052,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1e4981db-db53-4c47-be46-2f85b39c34a2
	{"id":"1e4981db-db53-4c47-be46-2f85b39c34a2","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822821}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43677052d4
	1e4981db-db53-4c47-be46-2f85b39c34a2

/tidb/cdc/default/default/upstream/7365107873580712052
	{"id":7365107873580712052,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.row_format.cli.2934.out cli changefeed create --start-ts=449530513271554049 '--sink-uri=kafka://127.0.0.1:9092/ticdc-row-format-test-5647?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760'
Create changefeed successfully!
ID: 4d8cff0d-7172-47d1-9cca-cbc1012bee8f
Info: {"upstream_id":7365107873580712052,"namespace":"default","id":"4d8cff0d-7172-47d1-9cca-cbc1012bee8f","sink_uri":"kafka://127.0.0.1:9092/ticdc-row-format-test-5647?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:40:24.802033323+08:00","start_ts":449530513271554049,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530513271554049,"checkpoint_ts":449530513271554049,"checkpoint_time":"2024-05-04 19:40:19.792"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
+ set +x
[Sat May  4 19:40:26 CST 2024] <<<<<< START kafka consumer in row_format case >>>>>>
table charset_gbk_test0.t0 exists
table charset_gbk_test0.t1 exists
table charset_gbk_test1.t0 not exists for 1-th check, retry later
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table charset_gbk_test1.t0 exists
table test.finish_mark not exists for 1-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table test.finish_mark not exists for 2-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 28.78 secs (129517904 bytes/sec)
[Pipeline] {
[Pipeline] cache
table test.finish_mark not exists for 3-th check, retry later
table foreign_key.finish_mark not exists for 1-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table foreign_key.finish_mark not exists for 2-th check, retry later
table test.finish_mark not exists for 4-th check, retry later
table row_format.finish_mark not exists for 1-th check, retry later
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9f9220000d	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:1844, start at 2024-05-04 19:40:33.299778493 +0800 CST m=+5.271051390	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:33.307 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:33.288 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:33.288 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9f9220000d	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:1844, start at 2024-05-04 19:40:33.299778493 +0800 CST m=+5.271051390	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:33.307 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:33.288 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:33.288 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0d9f92080014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:1920, start at 2024-05-04 19:40:33.304611907 +0800 CST m=+5.223872508	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:42:33.311 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:40:33.282 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:30:33.282 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/kafka_big_messages/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/kafka_big_messages/tiflash/log/error.log
arg matches is ArgMatches { args: {"config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/kafka_big_messages/tiflash-proxy.toml"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/kafka_big_messages/tiflash/log/proxy.log"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/kafka_big_messages/tiflash/db/proxy"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
table foreign_key.finish_mark not exists for 3-th check, retry later
table test.finish_mark not exists for 5-th check, retry later
table row_format.finish_mark not exists for 2-th check, retry later
[Sat May  4 19:40:38 CST 2024] <<<<<< START cdc server in kafka_big_messages case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ GO_FAILPOINTS=
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_big_messages.33713373.out server --log-file /tmp/tidb_cdc_test/kafka_big_messages/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/kafka_big_messages/cdc_data --cluster-id default
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
table foreign_key.finish_mark not exists for 4-th check, retry later
table test.finish_mark exists
check table exists success
check diff successfully
wait process cdc.test exit for 1-th time...
table row_format.finish_mark not exists for 3-th check, retry later
wait process cdc.test exit for 2-th time...
cdc.test: no process found
wait process cdc.test exit for 3-th time...
process cdc.test already exit
[Sat May  4 19:40:40 CST 2024] <<<<<< run test case charset_gbk success! >>>>>>
table foreign_key.finish_mark not exists for 5-th check, retry later
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:40:41 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/bf7302c0-f84a-4e0a-b346-7fcd16a84e2c
	{"id":"bf7302c0-f84a-4e0a-b346-7fcd16a84e2c","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822838}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4367be09cc
	bf7302c0-f84a-4e0a-b346-7fcd16a84e2c

/tidb/cdc/default/default/upstream/7365107949884048230
	{"id":7365107949884048230,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/bf7302c0-f84a-4e0a-b346-7fcd16a84e2c
	{"id":"bf7302c0-f84a-4e0a-b346-7fcd16a84e2c","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822838}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4367be09cc
	bf7302c0-f84a-4e0a-b346-7fcd16a84e2c

/tidb/cdc/default/default/upstream/7365107949884048230
	{"id":7365107949884048230,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/bf7302c0-f84a-4e0a-b346-7fcd16a84e2c
	{"id":"bf7302c0-f84a-4e0a-b346-7fcd16a84e2c","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822838}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4367be09cc
	bf7302c0-f84a-4e0a-b346-7fcd16a84e2c

/tidb/cdc/default/default/upstream/7365107949884048230
	{"id":7365107949884048230,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
Create changefeed successfully!
ID: 5f32bd8a-2070-4926-9a46-8e4f827bbac3
Info: {"upstream_id":7365107949884048230,"namespace":"default","id":"5f32bd8a-2070-4926-9a46-8e4f827bbac3","sink_uri":"kafka://127.0.0.1:9092/big-message-test?protocol=open-protocol\u0026partition-num=1\u0026kafka-version=2.4.1\u0026max-message-bytes=12582912","create_time":"2024-05-04T19:40:41.744165199+08:00","start_ts":449530518159228929,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530518159228929,"checkpoint_ts":449530518159228929,"checkpoint_time":"2024-05-04 19:40:38.437"}
[Sat May  4 19:40:41 CST 2024] <<<<<< START kafka consumer in kafka_big_messages case >>>>>>
Starting generate kafka big messages...
go: downloading github.com/pingcap/errors v0.11.5-0.20240318064555-6bd07397691f
go: downloading go.uber.org/atomic v1.11.0
table row_format.finish_mark not exists for 4-th check, retry later
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 7.18 secs (518985952 bytes/sec)
[Pipeline] {
[Pipeline] cache
table foreign_key.finish_mark not exists for 6-th check, retry later
table row_format.finish_mark not exists for 5-th check, retry later
table foreign_key.finish_mark not exists for 7-th check, retry later
table row_format.finish_mark not exists for 6-th check, retry later
table foreign_key.finish_mark not exists for 8-th check, retry later
table row_format.finish_mark not exists for 7-th check, retry later
table kafka_big_messages.test exists
check diff failed 1-th time, retry later
table foreign_key.finish_mark exists
check diff successfully
wait process cdc.test exit for 1-th time...
table row_format.finish_mark not exists for 8-th check, retry later
wait process cdc.test exit for 2-th time...
cdc.test: no process found
wait process cdc.test exit for 3-th time...
process cdc.test already exit
[Sat May  4 19:40:50 CST 2024] <<<<<< run test case foreign_key success! >>>>>>
check diff failed 2-th time, retry later
table row_format.finish_mark not exists for 9-th check, retry later
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/ddl_manager/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
check diff successfully
wait process cdc.test exit for 1-th time...
wait process cdc.test exit for 2-th time...
table row_format.finish_mark not exists for 10-th check, retry later
cdc.test: no process found
wait process cdc.test exit for 3-th time...
process cdc.test already exit
[Sat May  4 19:40:54 CST 2024] <<<<<< run test case kafka_big_messages success! >>>>>>
table row_format.finish_mark not exists for 11-th check, retry later
start tidb cluster in /tmp/tidb_cdc_test/ddl_manager
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
table row_format.finish_mark not exists for 12-th check, retry later
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
table row_format.finish_mark not exists for 13-th check, retry later
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table row_format.finish_mark not exists for 14-th check, retry later
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/ddl_puller_lag/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table row_format.finish_mark not exists for 15-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
start tidb cluster in /tmp/tidb_cdc_test/ddl_puller_lag
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
table row_format.finish_mark not exists for 16-th check, retry later
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/kafka_compression/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da19448000d	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs, pid:4183, start at 2024-05-04 19:41:06.206429552 +0800 CST m=+5.176184278	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:43:06.215 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:41:06.194 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:31:06.194 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Verifying downstream PD is started...
table row_format.finish_mark not exists for 17-th check, retry later
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 23.62 secs (157771789 bytes/sec)
[Pipeline] {
[Pipeline] cache
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
start tidb cluster in /tmp/tidb_cdc_test/kafka_compression
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da19448000d	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs, pid:4183, start at 2024-05-04 19:41:06.206429552 +0800 CST m=+5.176184278	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:43:06.215 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:41:06.194 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:31:06.194 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da195ac000a	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs, pid:4271, start at 2024-05-04 19:41:06.293899025 +0800 CST m=+5.208921793	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:43:06.300 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:41:06.283 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:31:06.283 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/ddl_manager/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/ddl_manager/tiflash/log/error.log
arg matches is ArgMatches { args: {"addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/ddl_manager/tiflash/db/proxy"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/ddl_manager/tiflash-proxy.toml"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/ddl_manager/tiflash/log/proxy.log"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
table row_format.finish_mark not exists for 18-th check, retry later
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
[Sat May  4 19:41:11 CST 2024] <<<<<< START cdc server in ddl_manager case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ GO_FAILPOINTS='github.com/pingcap/tiflow/cdc/owner/ExecuteDDLSlowly=return(true)'
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.ddl_manager.56025604.out server --log-file /tmp/tidb_cdc_test/ddl_manager/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/ddl_manager/cdc_data --cluster-id default
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
table row_format.finish_mark not exists for 19-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
table row_format.finish_mark not exists for 20-th check, retry later
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:41:14 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/b33c19aa-f976-4c3d-8684-405cefdd9901
	{"id":"b33c19aa-f976-4c3d-8684-405cefdd9901","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822871}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43683ef6cc
	b33c19aa-f976-4c3d-8684-405cefdd9901

/tidb/cdc/default/default/upstream/7365108091965360378
	{"id":7365108091965360378,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/b33c19aa-f976-4c3d-8684-405cefdd9901
	{"id":"b33c19aa-f976-4c3d-8684-405cefdd9901","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822871}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43683ef6cc
	b33c19aa-f976-4c3d-8684-405cefdd9901

/tidb/cdc/default/default/upstream/7365108091965360378
	{"id":7365108091965360378,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/b33c19aa-f976-4c3d-8684-405cefdd9901
	{"id":"b33c19aa-f976-4c3d-8684-405cefdd9901","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822871}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43683ef6cc
	b33c19aa-f976-4c3d-8684-405cefdd9901

/tidb/cdc/default/default/upstream/7365108091965360378
	{"id":7365108091965360378,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.ddl_manager.cli.5666.out cli changefeed create '--sink-uri=kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760' -c=ddl-manager
Create changefeed successfully!
ID: ddl-manager
Info: {"upstream_id":7365108091965360378,"namespace":"default","id":"ddl-manager","sink_uri":"kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:41:14.896873234+08:00","start_ts":449530527676628995,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530527676628995,"checkpoint_ts":449530527676628995,"checkpoint_time":"2024-05-04 19:41:14.743"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
Starting Upstream TiDB...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table row_format.finish_mark not exists for 21-th check, retry later
+ set +x
[Sat May  4 19:41:16 CST 2024] <<<<<< START kafka consumer in ddl_manager case >>>>>>
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da229a4000b	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv, pid:4137, start at 2024-05-04 19:41:15.763005523 +0800 CST m=+5.138120092	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:43:15.769 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:41:15.753 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:31:15.753 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da229a4000b	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv, pid:4137, start at 2024-05-04 19:41:15.763005523 +0800 CST m=+5.138120092	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:43:15.769 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:41:15.753 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:31:15.753 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da22a480014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-pfc68-l59tv, pid:4213, start at 2024-05-04 19:41:15.82327515 +0800 CST m=+5.146564713	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:43:15.830 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:41:15.794 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:31:15.794 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/ddl_puller_lag/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/ddl_puller_lag/tiflash/log/error.log
arg matches is ArgMatches { args: {"config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/ddl_puller_lag/tiflash-proxy.toml"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/ddl_puller_lag/tiflash/log/proxy.log"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/ddl_puller_lag/tiflash/db/proxy"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table row_format.finish_mark not exists for 22-th check, retry later
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.ddl_puller_lag.cli.5609.out cli tso query --pd=http://127.0.0.1:2379
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table row_format.finish_mark not exists for 23-th check, retry later
wait process 5607 exit for 1-th time...
wait process 5607 exit for 2-th time...
+ set +x
+ tso='449530528858636289
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530528858636289 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
[Sat May  4 19:41:20 CST 2024] <<<<<< START cdc server in ddl_puller_lag case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ GO_FAILPOINTS='github.com/pingcap/tiflow/cdc/processor/processorDDLResolved=1*sleep(180000)'
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.ddl_puller_lag.56565658.out server --log-file /tmp/tidb_cdc_test/ddl_puller_lag/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/ddl_puller_lag/cdc_data --cluster-id default
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
wait process 5607 exit for 3-th time...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da26f94001b	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:5839, start at 2024-05-04 19:41:20.27414998 +0800 CST m=+5.148982269	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:43:20.282 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:41:20.279 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:31:20.279 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da26f94001b	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:5839, start at 2024-05-04 19:41:20.27414998 +0800 CST m=+5.148982269	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:43:20.282 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:41:20.279 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:31:20.279 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da271180013	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:5917, start at 2024-05-04 19:41:20.347388102 +0800 CST m=+5.170067995	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:43:20.353 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:41:20.326 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:31:20.326 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/kafka_compression/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/kafka_compression/tiflash/log/error.log
arg matches is ArgMatches { args: {"config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/kafka_compression/tiflash-proxy.toml"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/kafka_compression/tiflash/log/proxy.log"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/kafka_compression/tiflash/db/proxy"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils/kill_cdc_pid: line 19: kill: (5607) - No such process
wait process 5607 exit for 4-th time...
process 5607 already exit
[Sat May  4 19:41:21 CST 2024] <<<<<< START cdc server in ddl_manager case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ GO_FAILPOINTS='github.com/pingcap/tiflow/cdc/owner/ExecuteDDLSlowly=return(true)'
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.ddl_manager.57705772.out server --log-file /tmp/tidb_cdc_test/ddl_manager/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/ddl_manager/cdc_data --cluster-id default
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
table row_format.finish_mark not exists for 24-th check, retry later
table row_format.finish_mark not exists for 25-th check, retry later
[Sat May  4 19:41:23 CST 2024] <<<<<< START cdc server in kafka_compression case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ GO_FAILPOINTS=
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.72397241.out server --log-file /tmp/tidb_cdc_test/kafka_compression/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/kafka_compression/cdc_data --cluster-id default
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:41:23 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/3dcf937a-8da4-48af-b75b-4e6e198006dc
	{"id":"3dcf937a-8da4-48af-b75b-4e6e198006dc","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822881}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43686067eb
	3dcf937a-8da4-48af-b75b-4e6e198006dc

/tidb/cdc/default/default/upstream/7365108131428423074
	{"id":7365108131428423074,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/3dcf937a-8da4-48af-b75b-4e6e198006dc
	{"id":"3dcf937a-8da4-48af-b75b-4e6e198006dc","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822881}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43686067eb
	3dcf937a-8da4-48af-b75b-4e6e198006dc

/tidb/cdc/default/default/upstream/7365108131428423074
	{"id":7365108131428423074,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/3dcf937a-8da4-48af-b75b-4e6e198006dc
	{"id":"3dcf937a-8da4-48af-b75b-4e6e198006dc","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822881}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43686067eb
	3dcf937a-8da4-48af-b75b-4e6e198006dc

/tidb/cdc/default/default/upstream/7365108131428423074
	{"id":7365108131428423074,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.ddl_puller_lag.cli.5709.out cli changefeed create --start-ts=449530528858636289 '--sink-uri=kafka+ssl://127.0.0.1:9092/ticdc-ddl-puller-lag-test-15022?protocol=open-protocol&partition-num=4&kafka-client-id=ddl_puller_lag&kafka-version=2.4.1&max-message-bytes=10485760'
Create changefeed successfully!
ID: fcb07783-46f8-4164-b7db-fb3f877eb687
Info: {"upstream_id":7365108131428423074,"namespace":"default","id":"fcb07783-46f8-4164-b7db-fb3f877eb687","sink_uri":"kafka+ssl://127.0.0.1:9092/ticdc-ddl-puller-lag-test-15022?protocol=open-protocol\u0026partition-num=4\u0026kafka-client-id=ddl_puller_lag\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:41:24.444911039+08:00","start_ts":449530528858636289,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530528858636289,"checkpoint_ts":449530528858636289,"checkpoint_time":"2024-05-04 19:41:19.252"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
+ set +x
[Sat May  4 19:41:25 CST 2024] <<<<<< START kafka consumer in ddl_puller_lag case >>>>>>
table row_format.finish_mark not exists for 26-th check, retry later
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:41:26 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1947b6f2-cc74-49bb-9459-8fe9cbc30b5f
	{"id":"1947b6f2-cc74-49bb-9459-8fe9cbc30b5f","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822883}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43687021c4
	1947b6f2-cc74-49bb-9459-8fe9cbc30b5f

/tidb/cdc/default/default/upstream/7365108155204314861
	{"id":7365108155204314861,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1947b6f2-cc74-49bb-9459-8fe9cbc30b5f
	{"id":"1947b6f2-cc74-49bb-9459-8fe9cbc30b5f","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822883}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43687021c4
	1947b6f2-cc74-49bb-9459-8fe9cbc30b5f

/tidb/cdc/default/default/upstream/7365108155204314861
	{"id":7365108155204314861,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1947b6f2-cc74-49bb-9459-8fe9cbc30b5f
	{"id":"1947b6f2-cc74-49bb-9459-8fe9cbc30b5f","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822883}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43687021c4
	1947b6f2-cc74-49bb-9459-8fe9cbc30b5f

/tidb/cdc/default/default/upstream/7365108155204314861
	{"id":7365108155204314861,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7301.out cli tso query --pd=http://127.0.0.1:2379
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:41:27 GMT
< Content-Type: text/plain; charset=utf-8
< Transfer-Encoding: chunked
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:

changefeedID: default/ddl-manager
{UpstreamID:7365108091965360378 Namespace:default ID:ddl-manager SinkURI:kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760 CreateTime:2024-05-04 19:41:14.896873234 +0800 CST StartTs:449530527676628995 TargetTs:0 AdminJobType:noop Engine:unified SortDir: Config:0xc00387ebd0 State:normal Error:<nil> Warning:<nil> CreatorVersion:v8.2.0-alpha-88-g57a3c8e82 Epoch:449530527716212737}
{CheckpointTs:449530528122273818 MinTableBarrierTs:449530528122273818 AdminJobType:noop}



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1e437d85-b189-4005-ad32-3d72df28252b
	{"id":"1e437d85-b189-4005-ad32-3d72df28252b","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822881}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43683ef94c
	1e437d85-b189-4005-ad32-3d72df28252b

/tidb/cdc/default/default/changefeed/info/ddl-manager
	{"upstream-id":7365108091965360378,"namespace":"default","changefeed-id":"ddl-manager","sink-uri":"kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create-time":"2024-05-04T19:41:14.896873234+08:00","start-ts":449530527676628995,"target-ts":0,"admin-job-type":0,"sort-engine":"","sort-dir":"","config":{"memory-quota":1073741824,"case-sensitive":false,"force-replicate":false,"check-gc-safe-point":true,"enable-sync-point":false,"enable-table-monitor":false,"ignore-ineligible-table":false,"bdr-mode":false,"sync-point-interval":600000000000,"sync-point-retention":86400000000000,"filter":{"rules":["*.*"],"ignore-txn-start-ts":null,"event-filters":null},"mounter":{"worker-num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include-commit-ts":false,"binary-encoding-method":"base64","output-old-value":false,"output-handle-key":false},"encoder-concurrency":32,"terminator":"\r\n","date-separator":"day","enable-partition-separator":true,"enable-kafka-sink-v2":false,"only-output-updated-columns":false,"delete-only-output-handle-key-columns":false,"content-compatible":false,"advance-timeout-in-sec":150,"send-bootstrap-interval-in-sec":120,"send-bootstrap-in-msg-count":10000,"send-bootstrap-to-all-partition":true,"debezium-disable-schema":false,"open":{"output-old-value":true},"debezium":{"output-old-value":true}},"consistent":{"level":"none","max-log-size":64,"flush-interval":2000,"meta-flush-interval":200,"encoding-worker-num":16,"flush-worker-num":8,"storage":"","use-file-backend":false,"compression":"","memory-usage":{"memory-quota-percentage":50}},"scheduler":{"enable-table-across-nodes":false,"region-threshold":100000,"write-key-threshold":0,"region-per-span":0},"integrity":{"integrity-check-level":"none","corruption-handle-level":"warn"},"changefeed-error-stuck-duration":1800000000000,"synced-status":{"synced-check-interval":300,"checkpoint-interval":15},"sql-mode":""},"state":"normal","error":null,"warning":null,"creator-version":"v8.2.0-alpha-88-g57a3c8e82","epoch":449530527716212737}

/tidb/cdc/default/default/changefeed/status/ddl-manager
	{"checkpoint-ts":449530528122273818,"min-table-barrier-ts":449530528122273818,"admin-job-type":0}

/tidb/cdc/default/default/task/position/1e437d85-b189-4005-ad32-3d72df28252b/ddl-manager
	{"checkpoint-ts":0,"resolved-ts":0,"count":0,"error":null,"warning":null}

/tidb/cdc/default/default/upstream/7365108091965360378
	{"id":7365108091965360378,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:

changefeedID: default/ddl-manager
{UpstreamID:7365108091965360378 Namespace:default ID:ddl-manager SinkURI:kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760 CreateTime:2024-05-04 19:41:14.896873234 +0800 CST StartTs:449530527676628995 TargetTs:0 AdminJobType:noop Engine:unified SortDir: Config:0xc00387ebd0 State:normal Error:<nil> Warning:<nil> CreatorVersion:v8.2.0-alpha-88-g57a3c8e82 Epoch:449530527716212737}
{CheckpointTs:449530528122273818 MinTableBarrierTs:449530528122273818 AdminJobType:noop}



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1e437d85-b189-4005-ad32-3d72df28252b
	{"id":"1e437d85-b189-4005-ad32-3d72df28252b","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822881}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43683ef94c
	1e437d85-b189-4005-ad32-3d72df28252b

/tidb/cdc/default/default/changefeed/info/ddl-manager
	{"upstream-id":7365108091965360378,"namespace":"default","changefeed-id":"ddl-manager","sink-uri":"kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create-time":"2024-05-04T19:41:14.896873234+08:00","start-ts":449530527676628995,"target-ts":0,"admin-job-type":0,"sort-engine":"","sort-dir":"","config":{"memory-quota":1073741824,"case-sensitive":false,"force-replicate":false,"check-gc-safe-point":true,"enable-sync-point":false,"enable-table-monitor":false,"ignore-ineligible-table":false,"bdr-mode":false,"sync-point-interval":600000000000,"sync-point-retention":86400000000000,"filter":{"rules":["*.*"],"ignore-txn-start-ts":null,"event-filters":null},"mounter":{"worker-num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include-commit-ts":false,"binary-encoding-method":"base64","output-old-value":false,"output-handle-key":false},"encoder-concurrency":32,"terminator":"\r\n","date-separator":"day","enable-partition-separator":true,"enable-kafka-sink-v2":false,"only-output-updated-columns":false,"delete-only-output-handle-key-columns":false,"content-compatible":false,"advance-timeout-in-sec":150,"send-bootstrap-interval-in-sec":120,"send-bootstrap-in-msg-count":10000,"send-bootstrap-to-all-partition":true,"debezium-disable-schema":false,"open":{"output-old-value":true},"debezium":{"output-old-value":true}},"consistent":{"level":"none","max-log-size":64,"flush-interval":2000,"meta-flush-interval":200,"encoding-worker-num":16,"flush-worker-num":8,"storage":"","use-file-backend":false,"compression":"","memory-usage":{"memory-quota-percentage":50}},"scheduler":{"enable-table-across-nodes":false,"region-threshold":100000,"write-key-threshold":0,"region-per-span":0},"integrity":{"integrity-check-level":"none","corruption-handle-level":"warn"},"changefeed-error-stuck-duration":1800000000000,"synced-status":{"synced-check-interval":300,"checkpoint-interval":15},"sql-mode":""},"state":"normal","error":null,"warning":null,"creator-version":"v8.2.0-alpha-88-g57a3c8e82","epoch":449530527716212737}

/tidb/cdc/default/default/changefeed/status/ddl-manager
	{"checkpoint-ts":449530528122273818,"min-table-barrier-ts":449530528122273818,"admin-job-type":0}

/tidb/cdc/default/default/task/position/1e437d85-b189-4005-ad32-3d72df28252b/ddl-manager
	{"checkpoint-ts":0,"resolved-ts":0,"count":0,"error":null,"warning":null}

/tidb/cdc/default/default/upstream/7365108091965360378
	{"id":7365108091965360378,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:

changefeedID: default/ddl-manager
{UpstreamID:7365108091965360378 Namespace:default ID:ddl-manager SinkURI:kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760 CreateTime:2024-05-04 19:41:14.896873234 +0800 CST StartTs:449530527676628995 TargetTs:0 AdminJobType:noop Engine:unified SortDir: Config:0xc00387ebd0 State:normal Error:<nil> Warning:<nil> CreatorVersion:v8.2.0-alpha-88-g57a3c8e82 Epoch:449530527716212737}
{CheckpointTs:449530528122273818 MinTableBarrierTs:449530528122273818 AdminJobType:noop}



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1e437d85-b189-4005-ad32-3d72df28252b
	{"id":"1e437d85-b189-4005-ad32-3d72df28252b","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822881}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43683ef94c
	1e437d85-b189-4005-ad32-3d72df28252b

/tidb/cdc/default/default/changefeed/info/ddl-manager
	{"upstream-id":7365108091965360378,"namespace":"default","changefeed-id":"ddl-manager","sink-uri":"kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create-time":"2024-05-04T19:41:14.896873234+08:00","start-ts":449530527676628995,"target-ts":0,"admin-job-type":0,"sort-engine":"","sort-dir":"","config":{"memory-quota":1073741824,"case-sensitive":false,"force-replicate":false,"check-gc-safe-point":true,"enable-sync-point":false,"enable-table-monitor":false,"ignore-ineligible-table":false,"bdr-mode":false,"sync-point-interval":600000000000,"sync-point-retention":86400000000000,"filter":{"rules":["*.*"],"ignore-txn-start-ts":null,"event-filters":null},"mounter":{"worker-num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include-commit-ts":false,"binary-encoding-method":"base64","output-old-value":false,"output-handle-key":false},"encoder-concurrency":32,"terminator":"\r\n","date-separator":"day","enable-partition-separator":true,"enable-kafka-sink-v2":false,"only-output-updated-columns":false,"delete-only-output-handle-key-columns":false,"content-compatible":false,"advance-timeout-in-sec":150,"send-bootstrap-interval-in-sec":120,"send-bootstrap-in-msg-count":10000,"send-bootstrap-to-all-partition":true,"debezium-disable-schema":false,"open":{"output-old-value":true},"debezium":{"output-old-value":true}},"consistent":{"level":"none","max-log-size":64,"flush-interval":2000,"meta-flush-interval":200,"encoding-worker-num":16,"flush-worker-num":8,"storage":"","use-file-backend":false,"compression":"","memory-usage":{"memory-quota-percentage":50}},"scheduler":{"enable-table-across-nodes":false,"region-threshold":100000,"write-key-threshold":0,"region-per-span":0},"integrity":{"integrity-check-level":"none","corruption-handle-level":"warn"},"changefeed-error-stuck-duration":1800000000000,"synced-status":{"synced-check-interval":300,"checkpoint-interval":15},"sql-mode":""},"state":"normal","error":null,"warning":null,"creator-version":"v8.2.0-alpha-88-g57a3c8e82","epoch":449530527716212737}

/tidb/cdc/default/default/changefeed/status/ddl-manager
	{"checkpoint-ts":449530528122273818,"min-table-barrier-ts":449530528122273818,"admin-job-type":0}

/tidb/cdc/default/default/task/position/1e437d85-b189-4005-ad32-3d72df28252b/ddl-manager
	{"checkpoint-ts":0,"resolved-ts":0,"count":0,"error":null,"warning":null}

/tidb/cdc/default/default/upstream/7365108091965360378
	{"id":7365108091965360378,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
[Sat May  4 19:41:27 CST 2024] <<<<<< START cdc server in ddl_manager case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ GO_FAILPOINTS='github.com/pingcap/tiflow/cdc/owner/ExecuteDDLSlowly=return(true)'
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.ddl_manager.58245826.out server --log-file /tmp/tidb_cdc_test/ddl_manager/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/ddl_manager/cdc_data --cluster-id default
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
table row_format.finish_mark not exists for 27-th check, retry later
+ set +x
+ tso='449530530844901380
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530530844901380 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7339.out cli changefeed create --start-ts=449530530844901380 '--sink-uri=kafka://127.0.0.1:9092/ticdc-kafka-compression-gzip-test?protocol=canal-json&enable-tidb-extension=true&kafka-version=2.4.1&compression=gzip' -c gzip
Create changefeed successfully!
ID: gzip
Info: {"upstream_id":7365108155204314861,"namespace":"default","id":"gzip","sink_uri":"kafka://127.0.0.1:9092/ticdc-kafka-compression-gzip-test?protocol=canal-json\u0026enable-tidb-extension=true\u0026kafka-version=2.4.1\u0026compression=gzip","create_time":"2024-05-04T19:41:28.720069977+08:00","start_ts":449530530844901380,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"canal-json","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530530844901380,"checkpoint_ts":449530530844901380,"checkpoint_time":"2024-05-04 19:41:26.829"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
table row_format.finish_mark not exists for 28-th check, retry later
+ set +x
[Sat May  4 19:41:30 CST 2024] <<<<<< START kafka consumer in kafka_compression case >>>>>>
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/kafka_compression/run.sh: line 22: [[: [2024/05/04 19:41:28.680 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]
[2024/05/04 19:41:28.715 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]
[2024/05/04 19:41:28.842 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]
[2024/05/04 19:41:28.853 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]
[2024/05/04 19:41:29.814 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]
[2024/05/04 19:41:29.823 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]: syntax error: operand expected (error token is "[2024/05/04 19:41:28.680 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]
[2024/05/04 19:41:28.715 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]
[2024/05/04 19:41:28.842 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]
[2024/05/04 19:41:28.853 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]
[2024/05/04 19:41:29.814 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]
[2024/05/04 19:41:29.823 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses gzip compression algorithm"]")
table test.gzip_finish_mark not exists for 1-th check, retry later
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:41:30 GMT
< Content-Type: text/plain; charset=utf-8
< Transfer-Encoding: chunked
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:

changefeedID: default/ddl-manager
{UpstreamID:7365108091965360378 Namespace:default ID:ddl-manager SinkURI:kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760 CreateTime:2024-05-04 19:41:14.896873234 +0800 CST StartTs:449530527676628995 TargetTs:0 AdminJobType:noop Engine:unified SortDir: Config:0xc00387ebd0 State:normal Error:<nil> Warning:<nil> CreatorVersion:v8.2.0-alpha-88-g57a3c8e82 Epoch:449530527716212737}
{CheckpointTs:449530528122273818 MinTableBarrierTs:449530528122273818 AdminJobType:noop}



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1e437d85-b189-4005-ad32-3d72df28252b
	{"id":"1e437d85-b189-4005-ad32-3d72df28252b","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822881}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43683ef94c
	1e437d85-b189-4005-ad32-3d72df28252b

/tidb/cdc/default/default/changefeed/info/ddl-manager
	{"upstream-id":7365108091965360378,"namespace":"default","changefeed-id":"ddl-manager","sink-uri":"kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create-time":"2024-05-04T19:41:14.896873234+08:00","start-ts":449530527676628995,"target-ts":0,"admin-job-type":0,"sort-engine":"","sort-dir":"","config":{"memory-quota":1073741824,"case-sensitive":false,"force-replicate":false,"check-gc-safe-point":true,"enable-sync-point":false,"enable-table-monitor":false,"ignore-ineligible-table":false,"bdr-mode":false,"sync-point-interval":600000000000,"sync-point-retention":86400000000000,"filter":{"rules":["*.*"],"ignore-txn-start-ts":null,"event-filters":null},"mounter":{"worker-num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include-commit-ts":false,"binary-encoding-method":"base64","output-old-value":false,"output-handle-key":false},"encoder-concurrency":32,"terminator":"\r\n","date-separator":"day","enable-partition-separator":true,"enable-kafka-sink-v2":false,"only-output-updated-columns":false,"delete-only-output-handle-key-columns":false,"content-compatible":false,"advance-timeout-in-sec":150,"send-bootstrap-interval-in-sec":120,"send-bootstrap-in-msg-count":10000,"send-bootstrap-to-all-partition":true,"debezium-disable-schema":false,"open":{"output-old-value":true},"debezium":{"output-old-value":true}},"consistent":{"level":"none","max-log-size":64,"flush-interval":2000,"meta-flush-interval":200,"encoding-worker-num":16,"flush-worker-num":8,"storage":"","use-file-backend":false,"compression":"","memory-usage":{"memory-quota-percentage":50}},"scheduler":{"enable-table-across-nodes":false,"region-threshold":100000,"write-key-threshold":0,"region-per-span":0},"integrity":{"integrity-check-level":"none","corruption-handle-level":"warn"},"changefeed-error-stuck-duration":1800000000000,"synced-status":{"synced-check-interval":300,"checkpoint-interval":15},"sql-mode":""},"state":"normal","error":null,"warning":null,"creator-version":"v8.2.0-alpha-88-g57a3c8e82","epoch":449530527716212737}

/tidb/cdc/default/default/changefeed/status/ddl-manager
	{"checkpoint-ts":449530528135381029,"min-table-barrier-ts":449530528135381029,"admin-job-type":0}

/tidb/cdc/default/default/task/position/1e437d85-b189-4005-ad32-3d72df28252b/ddl-manager
	{"checkpoint-ts":0,"resolved-ts":0,"count":0,"error":null,"warning":null}

/tidb/cdc/default/default/upstream/7365108091965360378
	{"id":7365108091965360378,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:

changefeedID: default/ddl-manager
{UpstreamID:7365108091965360378 Namespace:default ID:ddl-manager SinkURI:kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760 CreateTime:2024-05-04 19:41:14.896873234 +0800 CST StartTs:449530527676628995 TargetTs:0 AdminJobType:noop Engine:unified SortDir: Config:0xc00387ebd0 State:normal Error:<nil> Warning:<nil> CreatorVersion:v8.2.0-alpha-88-g57a3c8e82 Epoch:449530527716212737}
{CheckpointTs:449530528122273818 MinTableBarrierTs:449530528122273818 AdminJobType:noop}



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1e437d85-b189-4005-ad32-3d72df28252b
	{"id":"1e437d85-b189-4005-ad32-3d72df28252b","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822881}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43683ef94c
	1e437d85-b189-4005-ad32-3d72df28252b

/tidb/cdc/default/default/changefeed/info/ddl-manager
	{"upstream-id":7365108091965360378,"namespace":"default","changefeed-id":"ddl-manager","sink-uri":"kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create-time":"2024-05-04T19:41:14.896873234+08:00","start-ts":449530527676628995,"target-ts":0,"admin-job-type":0,"sort-engine":"","sort-dir":"","config":{"memory-quota":1073741824,"case-sensitive":false,"force-replicate":false,"check-gc-safe-point":true,"enable-sync-point":false,"enable-table-monitor":false,"ignore-ineligible-table":false,"bdr-mode":false,"sync-point-interval":600000000000,"sync-point-retention":86400000000000,"filter":{"rules":["*.*"],"ignore-txn-start-ts":null,"event-filters":null},"mounter":{"worker-num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include-commit-ts":false,"binary-encoding-method":"base64","output-old-value":false,"output-handle-key":false},"encoder-concurrency":32,"terminator":"\r\n","date-separator":"day","enable-partition-separator":true,"enable-kafka-sink-v2":false,"only-output-updated-columns":false,"delete-only-output-handle-key-columns":false,"content-compatible":false,"advance-timeout-in-sec":150,"send-bootstrap-interval-in-sec":120,"send-bootstrap-in-msg-count":10000,"send-bootstrap-to-all-partition":true,"debezium-disable-schema":false,"open":{"output-old-value":true},"debezium":{"output-old-value":true}},"consistent":{"level":"none","max-log-size":64,"flush-interval":2000,"meta-flush-interval":200,"encoding-worker-num":16,"flush-worker-num":8,"storage":"","use-file-backend":false,"compression":"","memory-usage":{"memory-quota-percentage":50}},"scheduler":{"enable-table-across-nodes":false,"region-threshold":100000,"write-key-threshold":0,"region-per-span":0},"integrity":{"integrity-check-level":"none","corruption-handle-level":"warn"},"changefeed-error-stuck-duration":1800000000000,"synced-status":{"synced-check-interval":300,"checkpoint-interval":15},"sql-mode":""},"state":"normal","error":null,"warning":null,"creator-version":"v8.2.0-alpha-88-g57a3c8e82","epoch":449530527716212737}

/tidb/cdc/default/default/changefeed/status/ddl-manager
	{"checkpoint-ts":449530528135381029,"min-table-barrier-ts":449530528135381029,"admin-job-type":0}

/tidb/cdc/default/default/task/position/1e437d85-b189-4005-ad32-3d72df28252b/ddl-manager
	{"checkpoint-ts":0,"resolved-ts":0,"count":0,"error":null,"warning":null}

/tidb/cdc/default/default/upstream/7365108091965360378
+ grep -q 'failed to get info:'
	{"id":7365108091965360378,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:

changefeedID: default/ddl-manager
{UpstreamID:7365108091965360378 Namespace:default ID:ddl-manager SinkURI:kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760 CreateTime:2024-05-04 19:41:14.896873234 +0800 CST StartTs:449530527676628995 TargetTs:0 AdminJobType:noop Engine:unified SortDir: Config:0xc00387ebd0 State:normal Error:<nil> Warning:<nil> CreatorVersion:v8.2.0-alpha-88-g57a3c8e82 Epoch:449530527716212737}
{CheckpointTs:449530528122273818 MinTableBarrierTs:449530528122273818 AdminJobType:noop}



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/1e437d85-b189-4005-ad32-3d72df28252b
	{"id":"1e437d85-b189-4005-ad32-3d72df28252b","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822881}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f43683ef94c
	1e437d85-b189-4005-ad32-3d72df28252b

/tidb/cdc/default/default/changefeed/info/ddl-manager
	{"upstream-id":7365108091965360378,"namespace":"default","changefeed-id":"ddl-manager","sink-uri":"kafka://127.0.0.1:9092/ticdc-ddl-mamager-test-1053?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create-time":"2024-05-04T19:41:14.896873234+08:00","start-ts":449530527676628995,"target-ts":0,"admin-job-type":0,"sort-engine":"","sort-dir":"","config":{"memory-quota":1073741824,"case-sensitive":false,"force-replicate":false,"check-gc-safe-point":true,"enable-sync-point":false,"enable-table-monitor":false,"ignore-ineligible-table":false,"bdr-mode":false,"sync-point-interval":600000000000,"sync-point-retention":86400000000000,"filter":{"rules":["*.*"],"ignore-txn-start-ts":null,"event-filters":null},"mounter":{"worker-num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include-commit-ts":false,"binary-encoding-method":"base64","output-old-value":false,"output-handle-key":false},"encoder-concurrency":32,"terminator":"\r\n","date-separator":"day","enable-partition-separator":true,"enable-kafka-sink-v2":false,"only-output-updated-columns":false,"delete-only-output-handle-key-columns":false,"content-compatible":false,"advance-timeout-in-sec":150,"send-bootstrap-interval-in-sec":120,"send-bootstrap-in-msg-count":10000,"send-bootstrap-to-all-partition":true,"debezium-disable-schema":false,"open":{"output-old-value":true},"debezium":{"output-old-value":true}},"consistent":{"level":"none","max-log-size":64,"flush-interval":2000,"meta-flush-interval":200,"encoding-worker-num":16,"flush-worker-num":8,"storage":"","use-file-backend":false,"compression":"","memory-usage":{"memory-quota-percentage":50}},"scheduler":{"enable-table-across-nodes":false,"region-threshold":100000,"write-key-threshold":0,"region-per-span":0},"integrity":{"integrity-check-level":"none","corruption-handle-level":"warn"},"changefeed-error-stuck-duration":1800000000000,"synced-status":{"synced-check-interval":300,"checkpoint-interval":15},"sql-mode":""},"state":"normal","error":null,"warning":null,"creator-version":"v8.2.0-alpha-88-g57a3c8e82","epoch":449530527716212737}

/tidb/cdc/default/default/changefeed/status/ddl-manager
	{"checkpoint-ts":449530528135381029,"min-table-barrier-ts":449530528135381029,"admin-job-type":0}

/tidb/cdc/default/default/task/position/1e437d85-b189-4005-ad32-3d72df28252b/ddl-manager
	{"checkpoint-ts":0,"resolved-ts":0,"count":0,"error":null,"warning":null}

/tidb/cdc/default/default/upstream/7365108091965360378
+ grep -q 'etcd info'
	{"id":7365108091965360378,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ break
+ set +x
table ddl_manager.finish_mark not exists for 1-th check, retry later
table row_format.finish_mark not exists for 29-th check, retry later
table ddl_manager.finish_mark not exists for 2-th check, retry later
table test.gzip_finish_mark not exists for 2-th check, retry later
table row_format.finish_mark not exists for 30-th check, retry later
table ddl_manager.finish_mark not exists for 3-th check, retry later
table test.gzip_finish_mark exists
check diff successfully
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7468.out cli changefeed pause -c gzip
PASS
coverage: 2.0% of statements in github.com/pingcap/tiflow/...
table row_format.finish_mark not exists for 31-th check, retry later
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7515.out cli changefeed remove -c gzip
table ddl_manager.finish_mark not exists for 4-th check, retry later
Changefeed remove successfully.
ID: gzip
CheckpointTs: 449530531762667550
SinkURI: kafka://127.0.0.1:9092/ticdc-kafka-compression-gzip-test?protocol=canal-json&enable-tidb-extension=true&kafka-version=2.4.1&compression=gzip
PASS
coverage: 2.0% of statements in github.com/pingcap/tiflow/...
table row_format.finish_mark not exists for 32-th check, retry later
table ddl_manager.finish_mark not exists for 5-th check, retry later
+ set +x
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7553.out cli tso query --pd=http://127.0.0.1:2379
table row_format.finish_mark not exists for 33-th check, retry later
table ddl_manager.finish_mark not exists for 6-th check, retry later
+ set +x
+ tso='449530534016843778
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530534016843778 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7588.out cli changefeed create --start-ts=449530534016843778 '--sink-uri=kafka://127.0.0.1:9092/ticdc-kafka-compression-snappy-test?protocol=canal-json&enable-tidb-extension=true&kafka-version=2.4.1&compression=snappy' -c snappy
Create changefeed successfully!
ID: snappy
Info: {"upstream_id":7365108155204314861,"namespace":"default","id":"snappy","sink_uri":"kafka://127.0.0.1:9092/ticdc-kafka-compression-snappy-test?protocol=canal-json\u0026enable-tidb-extension=true\u0026kafka-version=2.4.1\u0026compression=snappy","create_time":"2024-05-04T19:41:40.813521085+08:00","start_ts":449530534016843778,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"canal-json","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530534016843778,"checkpoint_ts":449530534016843778,"checkpoint_time":"2024-05-04 19:41:38.929"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
table row_format.finish_mark not exists for 34-th check, retry later
+ set +x
[Sat May  4 19:41:42 CST 2024] <<<<<< START kafka consumer in kafka_compression case >>>>>>
table ddl_manager.finish_mark not exists for 7-th check, retry later
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/kafka_compression/run.sh: line 22: [[: [2024/05/04 19:41:40.781 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]
[2024/05/04 19:41:40.809 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]
[2024/05/04 19:41:40.917 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]
[2024/05/04 19:41:40.925 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]
[2024/05/04 19:41:41.918 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]
[2024/05/04 19:41:41.926 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]: syntax error: operand expected (error token is "[2024/05/04 19:41:40.781 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]
[2024/05/04 19:41:40.809 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]
[2024/05/04 19:41:40.917 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]
[2024/05/04 19:41:40.925 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]
[2024/05/04 19:41:41.918 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]
[2024/05/04 19:41:41.926 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses snappy compression algorithm"]")
table test.snappy_finish_mark not exists for 1-th check, retry later
table row_format.finish_mark not exists for 35-th check, retry later
table test.snappy_finish_mark not exists for 2-th check, retry later
table ddl_manager.finish_mark not exists for 8-th check, retry later
table row_format.finish_mark not exists for 36-th check, retry later
table ddl_manager.finish_mark not exists for 9-th check, retry later
table test.snappy_finish_mark not exists for 3-th check, retry later
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 36.20 secs (102970665 bytes/sec)
[Pipeline] {
[Pipeline] cache
table row_format.finish_mark not exists for 37-th check, retry later
table test.snappy_finish_mark exists
check diff successfully
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7711.out cli changefeed pause -c snappy
table ddl_manager.finish_mark not exists for 10-th check, retry later
PASS
coverage: 2.0% of statements in github.com/pingcap/tiflow/...
table row_format.finish_mark not exists for 38-th check, retry later
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7747.out cli changefeed remove -c snappy
table ddl_manager.finish_mark not exists for 11-th check, retry later
Changefeed remove successfully.
ID: snappy
CheckpointTs: 449530536441937925
SinkURI: kafka://127.0.0.1:9092/ticdc-kafka-compression-snappy-test?protocol=canal-json&enable-tidb-extension=true&kafka-version=2.4.1&compression=snappy
PASS
coverage: 2.0% of statements in github.com/pingcap/tiflow/...
table row_format.finish_mark not exists for 39-th check, retry later
table ddl_manager.finish_mark not exists for 12-th check, retry later
+ set +x
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7781.out cli tso query --pd=http://127.0.0.1:2379
table row_format.finish_mark not exists for 40-th check, retry later
table ddl_manager.finish_mark not exists for 13-th check, retry later
+ set +x
+ tso='449530537687121923
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530537687121923 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7816.out cli changefeed create --start-ts=449530537687121923 '--sink-uri=kafka://127.0.0.1:9092/ticdc-kafka-compression-lz4-test?protocol=canal-json&enable-tidb-extension=true&kafka-version=2.4.1&compression=lz4' -c lz4
Create changefeed successfully!
ID: lz4
Info: {"upstream_id":7365108155204314861,"namespace":"default","id":"lz4","sink_uri":"kafka://127.0.0.1:9092/ticdc-kafka-compression-lz4-test?protocol=canal-json\u0026enable-tidb-extension=true\u0026kafka-version=2.4.1\u0026compression=lz4","create_time":"2024-05-04T19:41:54.816183029+08:00","start_ts":449530537687121923,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"canal-json","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530537687121923,"checkpoint_ts":449530537687121923,"checkpoint_time":"2024-05-04 19:41:52.930"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
table row_format.finish_mark not exists for 41-th check, retry later
+ set +x
[Sat May  4 19:41:56 CST 2024] <<<<<< START kafka consumer in kafka_compression case >>>>>>
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/kafka_compression/run.sh: line 22: [[: [2024/05/04 19:41:54.781 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]
[2024/05/04 19:41:54.811 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]
[2024/05/04 19:41:54.917 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]
[2024/05/04 19:41:54.925 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]
[2024/05/04 19:41:55.916 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]
[2024/05/04 19:41:55.925 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]: syntax error: operand expected (error token is "[2024/05/04 19:41:54.781 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]
[2024/05/04 19:41:54.811 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]
[2024/05/04 19:41:54.917 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]
[2024/05/04 19:41:54.925 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]
[2024/05/04 19:41:55.916 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]
[2024/05/04 19:41:55.925 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses lz4 compression algorithm"]")
table test.lz4_finish_mark not exists for 1-th check, retry later
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 7.03 secs (530548059 bytes/sec)
[Pipeline] {
[Pipeline] cache
table ddl_manager.finish_mark not exists for 14-th check, retry later
table row_format.finish_mark not exists for 42-th check, retry later
table test.lz4_finish_mark not exists for 2-th check, retry later
table ddl_manager.finish_mark not exists for 15-th check, retry later
table row_format.finish_mark not exists for 43-th check, retry later
table ddl_manager.finish_mark not exists for 16-th check, retry later
table test.lz4_finish_mark not exists for 3-th check, retry later
table row_format.finish_mark not exists for 44-th check, retry later
table ddl_manager.finish_mark not exists for 17-th check, retry later
table test.lz4_finish_mark exists
check diff successfully
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7963.out cli changefeed pause -c lz4
PASS
coverage: 2.0% of statements in github.com/pingcap/tiflow/...
table row_format.finish_mark not exists for 45-th check, retry later
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.7994.out cli changefeed remove -c lz4
table ddl_manager.finish_mark not exists for 18-th check, retry later
Changefeed remove successfully.
ID: lz4
CheckpointTs: 449530540111953926
SinkURI: kafka://127.0.0.1:9092/ticdc-kafka-compression-lz4-test?protocol=canal-json&enable-tidb-extension=true&kafka-version=2.4.1&compression=lz4
PASS
coverage: 2.0% of statements in github.com/pingcap/tiflow/...
table row_format.finish_mark not exists for 46-th check, retry later
+ set +x
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.8027.out cli tso query --pd=http://127.0.0.1:2379
table ddl_manager.finish_mark not exists for 19-th check, retry later
table row_format.finish_mark not exists for 47-th check, retry later
+ set +x
+ tso='449530541357137923
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530541357137923 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.8062.out cli changefeed create --start-ts=449530541357137923 '--sink-uri=kafka://127.0.0.1:9092/ticdc-kafka-compression-zstd-test?protocol=canal-json&enable-tidb-extension=true&kafka-version=2.4.1&compression=zstd' -c zstd
Create changefeed successfully!
ID: zstd
Info: {"upstream_id":7365108155204314861,"namespace":"default","id":"zstd","sink_uri":"kafka://127.0.0.1:9092/ticdc-kafka-compression-zstd-test?protocol=canal-json\u0026enable-tidb-extension=true\u0026kafka-version=2.4.1\u0026compression=zstd","create_time":"2024-05-04T19:42:08.818419984+08:00","start_ts":449530541357137923,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"canal-json","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530541357137923,"checkpoint_ts":449530541357137923,"checkpoint_time":"2024-05-04 19:42:06.930"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
table ddl_manager.finish_mark not exists for 20-th check, retry later
+ set +x
[Sat May  4 19:42:10 CST 2024] <<<<<< START kafka consumer in kafka_compression case >>>>>>
table row_format.finish_mark not exists for 48-th check, retry later
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/kafka_compression/run.sh: line 22: [[: [2024/05/04 19:42:08.784 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]
[2024/05/04 19:42:08.814 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]
[2024/05/04 19:42:08.919 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]
[2024/05/04 19:42:08.926 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]
[2024/05/04 19:42:09.918 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]
[2024/05/04 19:42:09.925 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]: syntax error: operand expected (error token is "[2024/05/04 19:42:08.784 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]
[2024/05/04 19:42:08.814 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]
[2024/05/04 19:42:08.919 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]
[2024/05/04 19:42:08.926 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]
[2024/05/04 19:42:09.918 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]
[2024/05/04 19:42:09.925 +08:00] [INFO] [sarama.go:96] ["Kafka producer uses zstd compression algorithm"]")
table test.zstd_finish_mark not exists for 1-th check, retry later
table ddl_manager.finish_mark not exists for 21-th check, retry later
table row_format.finish_mark not exists for 49-th check, retry later
table test.zstd_finish_mark not exists for 2-th check, retry later
table ddl_manager.finish_mark not exists for 22-th check, retry later
table row_format.finish_mark not exists for 50-th check, retry later
table ddl_manager.finish_mark not exists for 23-th check, retry later
table test.zstd_finish_mark not exists for 3-th check, retry later
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 16.80 secs (221793063 bytes/sec)
[Pipeline] {
[Pipeline] cache
table row_format.finish_mark not exists for 51-th check, retry later
table test.zstd_finish_mark exists
check diff successfully
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.8180.out cli changefeed pause -c zstd
table ddl_manager.finish_mark not exists for 24-th check, retry later
PASS
coverage: 2.0% of statements in github.com/pingcap/tiflow/...
table row_format.finish_mark not exists for 52-th check, retry later
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_compression.cli.8210.out cli changefeed remove -c zstd
table ddl_manager.finish_mark not exists for 25-th check, retry later
Changefeed remove successfully.
ID: zstd
CheckpointTs: 449530543795077123
SinkURI: kafka://127.0.0.1:9092/ticdc-kafka-compression-zstd-test?protocol=canal-json&enable-tidb-extension=true&kafka-version=2.4.1&compression=zstd
PASS
coverage: 2.0% of statements in github.com/pingcap/tiflow/...
table row_format.finish_mark not exists for 53-th check, retry later
+ set +x
table ddl_manager.finish_mark not exists for 26-th check, retry later
wait process cdc.test exit for 1-th time...
wait process cdc.test exit for 2-th time...
cdc.test: no process found
wait process cdc.test exit for 3-th time...
process cdc.test already exit
[Sat May  4 19:42:22 CST 2024] <<<<<< run test case kafka_compression success! >>>>>>
table row_format.finish_mark not exists for 54-th check, retry later
table ddl_manager.finish_mark not exists for 27-th check, retry later
table row_format.finish_mark not exists for 55-th check, retry later
table ddl_manager.finish_mark not exists for 28-th check, retry later
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 7.99 secs (466764083 bytes/sec)
[Pipeline] {
table row_format.finish_mark not exists for 56-th check, retry later
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] sh
table ddl_manager.finish_mark not exists for 29-th check, retry later
[Pipeline] sh
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
[Pipeline] sh
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
[Pipeline] sh
table row_format.finish_mark not exists for 57-th check, retry later
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ + grepnc brokers
 localhost 2181
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
+ echo dump
/brokers/ids/1
[Pipeline] sh
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
[Pipeline] sh
table ddl_manager.finish_mark not exists for 30-th check, retry later
[Pipeline] sh
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
table row_format.finish_mark not exists for 58-th check, retry later
table ddl_manager.finish_mark not exists for 31-th check, retry later
[Pipeline] archiveArtifacts
Archiving artifacts
Recording fingerprints
table row_format.finish_mark not exists for 59-th check, retry later
table ddl_manager.finish_mark not exists for 32-th check, retry later
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/kafka_messages/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:42:33 CST 2024] <<<<<< run test case kafka_messages success! >>>>>>
table row_format.finish_mark not exists for 60-th check, retry later
table ddl_manager.finish_mark not exists for 33-th check, retry later
table row_format.finish_mark not exists at last check
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/kafka_sink_error_resume/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
table ddl_manager.finish_mark not exists for 34-th check, retry later
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] }
[Pipeline] }
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 45 min
[Pipeline] {
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] withCredentials
Masking supported pattern matches of $TICDC_COVERALLS_TOKEN or $TICDC_CODECOV_TOKEN
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] }
table ddl_manager.finish_mark not exists for 35-th check, retry later
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] {
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] sh
[Pipeline] sh
start tidb cluster in /tmp/tidb_cdc_test/kafka_sink_error_resume
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G02
Run cases: consistent_replicate_ddl consistent_replicate_gbk consistent_replicate_nfs consistent_replicate_storage_file consistent_replicate_storage_file_large_value consistent_replicate_storage_s3 consistent_partition_table kafka_big_messages_v2 multi_tables_ddl_v2 multi_topics_v2 storage_cleanup csv_storage_basic csv_storage_multi_tables_ddl csv_storage_partition_table
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=620f62fe-a9f4-4fbb-b12f-f5a5237d9708
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G02
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-xx2km
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44 pingcap_tiflow_pull_cdc_integration_kafka_test_1845-xx2km
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/consistent_replicate_ddl/run.sh using Sink-Type: kafka... <<=================
[Pipeline] sh
[Sat May  4 19:42:40 CST 2024] <<<<<< run test case consistent_replicate_ddl success! >>>>>>
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)

+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G10
Run cases: default_value simple cdc_server_tips event_filter sql_mode
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=0b42d2bc-fa0c-49e9-85e3-a3998ce0860e
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G10
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-m43dx
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-m43dx pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/default_value/run.sh using Sink-Type: kafka... <<=================
[Pipeline] sh
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G11
Run cases: resolve_lock move_table autorandom generate_column
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=cab96a40-ed0d-4635-942f-e59251fa619c
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G11
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-ftwvg
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct pingcap_tiflow_pull_cdc_integration_kafka_test_1845-ftwvg
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/resolve_lock/run.sh using Sink-Type: kafka... <<=================
[Pipeline] sh
table ddl_manager.finish_mark not exists for 36-th check, retry later
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G08
Run cases: processor_err_chan changefeed_reconstruct multi_capture synced_status_with_redo
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=f72aa9f0-28bd-4765-aa73-22a3333be08c
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT=tcp://10.233.0.1:443
KUBERNETES_PORT_443_TCP_PORT=443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G08
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-t0ptc
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf pingcap_tiflow_pull_cdc_integration_kafka_test_1845-t0ptc
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/processor_err_chan/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
[Pipeline] sh
[Pipeline] // container
[Pipeline] sh
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G07
Run cases: kv_client_stream_reconnect cdc split_region
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=30faa295-a172-4a33-818e-4914bc86724f
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G07
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-6tpkp
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb pingcap_tiflow_pull_cdc_integration_kafka_test_1845-6tpkp
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/kv_client_stream_reconnect/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
[Pipeline] // timeout
[Pipeline] }
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G06
Run cases: sink_retry changefeed_error ddl_sequence resourcecontrol
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=c36397e8-d068-4f32-b540-dd4028846de1
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G06
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-321md
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd pingcap_tiflow_pull_cdc_integration_kafka_test_1845-321md
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/sink_retry/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
[Pipeline] cache
Verifying downstream PD is started...
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G09
Run cases: gc_safepoint changefeed_pause_resume cli_with_auth savepoint synced_status
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=db405ff9-68a9-4561-8459-f8a3cd041fba
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G09
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-gzt73
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk pingcap_tiflow_pull_cdc_integration_kafka_test_1845-gzt73
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/gc_safepoint/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
table ddl_manager.finish_mark not exists for 37-th check, retry later
The 1 times to try to start tidb cluster...
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/consistent_replicate_gbk/run.sh using Sink-Type: kafka... <<=================
* About to connect() to 127.0.0.1 port 24927 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:24927; Connection refused
* Closing connection 0
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release

 You are running an older version of MinIO released 3 years ago 
 Update: Run `mc admin update` 


Attempting encryption of all config, IAM users and policies on MinIO backend
start tidb cluster in /tmp/tidb_cdc_test/kv_client_stream_reconnect
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
The 1 times to try to start tidb cluster...
Endpoint:  http://127.0.0.1:24927

Object API (Amazon S3 compatible):
   Go:         https://docs.min.io/docs/golang-client-quickstart-guide
   Java:       https://docs.min.io/docs/java-client-quickstart-guide
   Python:     https://docs.min.io/docs/python-client-quickstart-guide
   JavaScript: https://docs.min.io/docs/javascript-client-quickstart-guide
   .NET:       https://docs.min.io/docs/dotnet-client-quickstart-guide
table ddl_manager.finish_mark not exists for 38-th check, retry later
start tidb cluster in /tmp/tidb_cdc_test/processor_err_chan
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
Starting Upstream TiDB...
* About to connect() to 127.0.0.1 port 24927 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 24927 (#0)
> GET / HTTP/1.1
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:24927
> Accept: */*
> 
< HTTP/1.1 403 Forbidden
< Accept-Ranges: bytes
< Content-Length: 226
< Content-Security-Policy: block-all-mixed-content
< Content-Type: application/xml
< Server: MinIO/RELEASE.2020-07-27T18-37-02Z
< Vary: Origin
< X-Amz-Request-Id: 17CC4666E409E747
< X-Xss-Protection: 1; mode=block
< Date: Sat, 04 May 2024 11:42:45 GMT
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
Bucket 's3://logbucket/' created
[Sat May  4 19:42:45 CST 2024] <<<<<< run test case consistent_replicate_gbk success! >>>>>>
Exiting on signal: INTERRUPT
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
start tidb cluster in /tmp/tidb_cdc_test/sink_retry
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
start tidb cluster in /tmp/tidb_cdc_test/default_value
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
start tidb cluster in /tmp/tidb_cdc_test/gc_safepoint
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
Verifying downstream PD is started...
table ddl_manager.finish_mark not exists for 39-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
start tidb cluster in /tmp/tidb_cdc_test/resolve_lock
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
table ddl_manager.finish_mark not exists for 40-th check, retry later
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/consistent_replicate_nfs/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:42:49 CST 2024] <<<<<< run test case consistent_replicate_nfs success! >>>>>>
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)

Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
table ddl_manager.finish_mark not exists for 41-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Verifying downstream PD is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da7fab80017	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:9394, start at 2024-05-04 19:42:51.103785435 +0800 CST m=+5.123454642	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:51.111 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:51.103 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:51.103 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da7fab80017	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:9394, start at 2024-05-04 19:42:51.103785435 +0800 CST m=+5.123454642	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:51.111 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:51.103 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:51.103 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da7fb7c0013	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:9481, start at 2024-05-04 19:42:51.128594041 +0800 CST m=+5.095587715	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:51.137 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:51.103 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:51.103 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/kafka_sink_error_resume/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/kafka_sink_error_resume/tiflash/log/error.log
arg matches is ArgMatches { args: {"engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/kafka_sink_error_resume/tiflash-proxy.toml"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/kafka_sink_error_resume/tiflash/db/proxy"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/kafka_sink_error_resume/tiflash/log/proxy.log"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Starting Upstream TiDB...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/consistent_replicate_storage_file/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:42:53 CST 2024] <<<<<< run test case consistent_replicate_storage_file success! >>>>>>
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)

table ddl_manager.finish_mark not exists for 42-th check, retry later
[Sat May  4 19:42:54 CST 2024] <<<<<< START cdc server in kafka_sink_error_resume case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ GO_FAILPOINTS='github.com/pingcap/tiflow/cdc/sink/dmlsink/mq/dmlproducer/KafkaSinkAsyncSendError=1*return(true)'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_sink_error_resume.1085810860.out server --log-file /tmp/tidb_cdc_test/kafka_sink_error_resume/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/kafka_sink_error_resume/cdc_data --cluster-id default --addr 127.0.0.1:8300 --pd http://127.0.0.1:2379
+ [[ no != \n\o ]]
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table ddl_manager.finish_mark not exists for 43-th check, retry later
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da8401c0010	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf, pid:1353, start at 2024-05-04 19:42:55.519386278 +0800 CST m=+5.229924402	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:55.529 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:55.495 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:55.495 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da842380016	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb, pid:1362, start at 2024-05-04 19:42:55.666681264 +0800 CST m=+5.057804368	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:55.672 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:55.630 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:55.630 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da842380016	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb, pid:1362, start at 2024-05-04 19:42:55.666681264 +0800 CST m=+5.057804368	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:55.672 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:55.630 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:55.630 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da845480014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb, pid:1441, start at 2024-05-04 19:42:55.852992819 +0800 CST m=+5.189449928	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:55.861 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:55.826 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:55.826 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/kv_client_stream_reconnect/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/kv_client_stream_reconnect/tiflash/log/error.log
arg matches is ArgMatches { args: {"engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/kv_client_stream_reconnect/tiflash/db/proxy"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/kv_client_stream_reconnect/tiflash/log/proxy.log"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/kv_client_stream_reconnect/tiflash-proxy.toml"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/consistent_replicate_storage_file_large_value/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:42:56 CST 2024] <<<<<< run test case consistent_replicate_storage_file_large_value success! >>>>>>
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)

+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:42:57 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/a26381c8-c293-4a96-9df4-8b813f1caa89
	{"id":"a26381c8-c293-4a96-9df4-8b813f1caa89","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822974}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369d2e9cb
	a26381c8-c293-4a96-9df4-8b813f1caa89

/tidb/cdc/default/default/upstream/7365108536470800469
	{"id":7365108536470800469,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/a26381c8-c293-4a96-9df4-8b813f1caa89
	{"id":"a26381c8-c293-4a96-9df4-8b813f1caa89","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822974}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369d2e9cb
	a26381c8-c293-4a96-9df4-8b813f1caa89

/tidb/cdc/default/default/upstream/7365108536470800469
	{"id":7365108536470800469,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/a26381c8-c293-4a96-9df4-8b813f1caa89
	{"id":"a26381c8-c293-4a96-9df4-8b813f1caa89","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822974}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369d2e9cb
	a26381c8-c293-4a96-9df4-8b813f1caa89

/tidb/cdc/default/default/upstream/7365108536470800469
	{"id":7365108536470800469,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
[Sat May  4 19:42:57 CST 2024] <<<<<< START kafka consumer in kafka_sink_error_resume case >>>>>>
table ddl_manager.finish_mark not exists for 44-th check, retry later
check_changefeed_status 127.0.0.1:8300 c5073bcc-4f80-43b5-8b38-d5bcab5984e5 warning last_warning kafka sink injected error
+ endpoint=127.0.0.1:8300
+ changefeed_id=c5073bcc-4f80-43b5-8b38-d5bcab5984e5
+ expected_state=warning
+ field=last_warning
+ error_pattern=kafka
++ curl 127.0.0.1:8300/api/v2/changefeeds/c5073bcc-4f80-43b5-8b38-d5bcab5984e5/status
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    86  100    86    0     0    600      0 --:--:-- --:--:-- --:--:--   601
+ info='{"state":"normal","resolved_ts":449530554601701377,"checkpoint_ts":449530554601701377}'
+ echo '{"state":"normal","resolved_ts":449530554601701377,"checkpoint_ts":449530554601701377}'
{"state":"normal","resolved_ts":449530554601701377,"checkpoint_ts":449530554601701377}
++ echo '{"state":"normal","resolved_ts":449530554601701377,"checkpoint_ts":449530554601701377}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \w\a\r\n\i\n\g ]]
+ echo 'changefeed state normal does not equal to warning'
changefeed state normal does not equal to warning
+ exit 1
run task failed 1-th time, retry later
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da8401c0010	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf, pid:1353, start at 2024-05-04 19:42:55.519386278 +0800 CST m=+5.229924402	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:55.529 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:55.495 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:55.495 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da8402c0015	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf, pid:1440, start at 2024-05-04 19:42:55.54280252 +0800 CST m=+5.197625855	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:55.551 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:55.549 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:55.549 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/processor_err_chan/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/processor_err_chan/tiflash/log/error.log
arg matches is ArgMatches { args: {"addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/processor_err_chan/tiflash/db/proxy"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/processor_err_chan/tiflash/log/proxy.log"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/processor_err_chan/tiflash-proxy.toml"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da861480006	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq, pid:1417, start at 2024-05-04 19:42:57.621883509 +0800 CST m=+5.113729007	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:57.630 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
[Sat May  4 19:42:58 CST 2024] <<<<<< START cdc server in kv_client_stream_reconnect case >>>>>>
tikv_gc_last_run_time	20240504-19:42:57.618 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:57.618 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da861480006	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq, pid:1417, start at 2024-05-04 19:42:57.621883509 +0800 CST m=+5.113729007	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:57.630 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:57.618 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:57.618 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da8639c000d	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-m43dx-g0tbq, pid:1507, start at 2024-05-04 19:42:57.780010699 +0800 CST m=+5.205237159	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:57.786 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:57.767 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:57.767 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/default_value/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/default_value/tiflash/log/error.log
arg matches is ArgMatches { args: {"log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/default_value/tiflash/log/proxy.log"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/default_value/tiflash/db/proxy"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/default_value/tiflash-proxy.toml"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ GO_FAILPOINTS='github.com/pingcap/tiflow/cdc/kv/kvClientForceReconnect=return(true)'
+ (( i = 0 ))
+ (( i <= 50 ))
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kv_client_stream_reconnect.28182820.out server --log-file /tmp/tidb_cdc_test/kv_client_stream_reconnect/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/kv_client_stream_reconnect/cdc_data --cluster-id default --addr 127.0.0.1:8300 --pd http://127.0.0.1:2379
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da86514000e	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk, pid:1356, start at 2024-05-04 19:42:57.876679468 +0800 CST m=+5.073120975	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:57.883 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:57.861 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:57.861 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da86514000e	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk, pid:1356, start at 2024-05-04 19:42:57.876679468 +0800 CST m=+5.073120975	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:57.883 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:57.861 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:57.861 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da867600014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk, pid:1442, start at 2024-05-04 19:42:58.027524198 +0800 CST m=+5.170258825	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:58.033 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:58.008 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:58.008 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/gc_safepoint/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/gc_safepoint/tiflash/log/error.log
arg matches is ArgMatches { args: {"advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/gc_safepoint/tiflash/db/proxy"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/gc_safepoint/tiflash/log/proxy.log"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/gc_safepoint/tiflash-proxy.toml"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da8482c0013	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd, pid:1296, start at 2024-05-04 19:42:56.052001483 +0800 CST m=+5.247237516	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:56.059 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:56.061 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:56.061 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da8482c0013	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd, pid:1296, start at 2024-05-04 19:42:56.052001483 +0800 CST m=+5.247237516	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:56.059 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:56.061 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:56.061 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da849b40014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd, pid:1369, start at 2024-05-04 19:42:56.13380084 +0800 CST m=+5.277440951	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:44:56.140 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:42:56.109 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:32:56.109 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/sink_retry/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/sink_retry/tiflash/log/error.log
arg matches is ArgMatches { args: {"log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/sink_retry/tiflash/log/proxy.log"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/sink_retry/tiflash/db/proxy"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/sink_retry/tiflash-proxy.toml"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
table ddl_manager.finish_mark not exists for 45-th check, retry later
check_changefeed_status 127.0.0.1:8300 c5073bcc-4f80-43b5-8b38-d5bcab5984e5 warning last_warning kafka sink injected error
+ endpoint=127.0.0.1:8300
+ changefeed_id=c5073bcc-4f80-43b5-8b38-d5bcab5984e5
+ expected_state=warning
+ field=last_warning
+ error_pattern=kafka
++ curl 127.0.0.1:8300/api/v2/changefeeds/c5073bcc-4f80-43b5-8b38-d5bcab5984e5/status
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    86  100    86    0     0    777      0 --:--:-- --:--:-- --:--:--   781
+ info='{"state":"normal","resolved_ts":449530554667237408,"checkpoint_ts":449530554667237408}'
+ echo '{"state":"normal","resolved_ts":449530554667237408,"checkpoint_ts":449530554667237408}'
{"state":"normal","resolved_ts":449530554667237408,"checkpoint_ts":449530554667237408}
++ echo '{"state":"normal","resolved_ts":449530554667237408,"checkpoint_ts":449530554667237408}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \w\a\r\n\i\n\g ]]
+ echo 'changefeed state normal does not equal to warning'
changefeed state normal does not equal to warning
+ exit 1
run task failed 2-th time, retry later
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/consistent_replicate_storage_s3/run.sh using Sink-Type: kafka... <<=================
* About to connect() to 127.0.0.1 port 24927 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:24927; Connection refused
* Closing connection 0

 You are running an older version of MinIO released 3 years ago 
 Update: Run `mc admin update` 


Attempting encryption of all config, IAM users and policies on MinIO backend
[Sat May  4 19:43:01 CST 2024] <<<<<< START cdc server in gc_safepoint case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ GO_FAILPOINTS='github.com/pingcap/tiflow/pkg/txnutil/gc/InjectGcSafepointUpdateInterval=return(500)'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.gc_safepoint.28252827.out server --log-file /tmp/tidb_cdc_test/gc_safepoint/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/gc_safepoint/cdc_data --cluster-id default --addr 127.0.0.1:8300 --pd http://127.0.0.1:2379
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da886b40013	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct, pid:1373, start at 2024-05-04 19:43:00.039494237 +0800 CST m=+5.044822098	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:00.045 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:00.013 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:00.013 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da886b40013	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct, pid:1373, start at 2024-05-04 19:43:00.039494237 +0800 CST m=+5.044822098	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:00.045 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:00.013 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:00.013 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da889080014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-ftwvg-pv8ct, pid:1451, start at 2024-05-04 19:43:00.200245517 +0800 CST m=+5.154000750	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:00.207 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:00.212 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:00.212 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.sink_retry.cli.2831.out cli tso query --pd=http://127.0.0.1:2379
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.default_value.cli.2902.out cli tso query --pd=http://127.0.0.1:2379
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/resolve_lock/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/resolve_lock/tiflash/log/error.log
arg matches is ArgMatches { args: {"engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/resolve_lock/tiflash/log/proxy.log"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/resolve_lock/tiflash/db/proxy"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/resolve_lock/tiflash-proxy.toml"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
Endpoint:  http://127.0.0.1:24927

Object API (Amazon S3 compatible):
   Go:         https://docs.min.io/docs/golang-client-quickstart-guide
   Java:       https://docs.min.io/docs/java-client-quickstart-guide
   Python:     https://docs.min.io/docs/python-client-quickstart-guide
   JavaScript: https://docs.min.io/docs/javascript-client-quickstart-guide
   .NET:       https://docs.min.io/docs/dotnet-client-quickstart-guide
table ddl_manager.finish_mark not exists for 46-th check, retry later
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 17.67 secs (210903504 bytes/sec)
[Pipeline] {
[Pipeline] cache
* About to connect() to 127.0.0.1 port 24927 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 24927 (#0)
> GET / HTTP/1.1
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:24927
> Accept: */*
> 
< HTTP/1.1 403 Forbidden
< Accept-Ranges: bytes
< Content-Length: 226
< Content-Security-Policy: block-all-mixed-content
< Content-Type: application/xml
< Server: MinIO/RELEASE.2020-07-27T18-37-02Z
< Vary: Origin
< X-Amz-Request-Id: 17CC466AB75FAFCD
< X-Xss-Protection: 1; mode=block
< Date: Sat, 04 May 2024 11:43:02 GMT
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:43:02 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/546e94a4-bb15-4d4d-b38e-3c54d5c79da9
	{"id":"546e94a4-bb15-4d4d-b38e-3c54d5c79da9","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822979}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369e6f0cb
	546e94a4-bb15-4d4d-b38e-3c54d5c79da9

/tidb/cdc/default/default/upstream/7365108553796313409
	{"id":7365108553796313409,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/546e94a4-bb15-4d4d-b38e-3c54d5c79da9
	{"id":"546e94a4-bb15-4d4d-b38e-3c54d5c79da9","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822979}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369e6f0cb
	546e94a4-bb15-4d4d-b38e-3c54d5c79da9

/tidb/cdc/default/default/upstream/7365108553796313409
	{"id":7365108553796313409,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/546e94a4-bb15-4d4d-b38e-3c54d5c79da9
	{"id":"546e94a4-bb15-4d4d-b38e-3c54d5c79da9","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822979}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369e6f0cb
	546e94a4-bb15-4d4d-b38e-3c54d5c79da9

/tidb/cdc/default/default/upstream/7365108553796313409
	{"id":7365108553796313409,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
[Sat May  4 19:43:02 CST 2024] <<<<<< START kafka consumer in kv_client_stream_reconnect case >>>>>>
Bucket 's3://logbucket/' created
[Sat May  4 19:43:02 CST 2024] <<<<<< run test case consistent_replicate_storage_s3 success! >>>>>>
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)

Exiting on signal: INTERRUPT
+ set +x
+ tso='449530555575304193
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530555575304193 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
[Sat May  4 19:43:02 CST 2024] <<<<<< START cdc server in default_value case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ GO_FAILPOINTS=
+ (( i = 0 ))
+ (( i <= 50 ))
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.default_value.29372939.out server --log-file /tmp/tidb_cdc_test/default_value/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/default_value/cdc_data --cluster-id default
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
[Sat May  4 19:43:02 CST 2024] <<<<<< START cdc server in processor_err_chan case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ GO_FAILPOINTS='github.com/pingcap/tiflow/cdc/processor/ProcessorAddTableError=1*return(true)'
+ (( i = 0 ))
+ (( i <= 50 ))
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.processor_err_chan.29692971.out server --log-file /tmp/tidb_cdc_test/processor_err_chan/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/processor_err_chan/cdc_data --cluster-id default --config /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/processor_err_chan/conf/server.toml --addr 127.0.0.1:8300 --pd http://127.0.0.1:2379
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
+ set +x
+ tso='449530555652112385
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530555652112385 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
table ddl_manager.finish_mark not exists for 47-th check, retry later
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.resolve_lock.cli.2888.out cli tso query --pd=http://127.0.0.1:2379
***************** properties *****************
"readproportion"="0"
"threadcount"="2"
"operationcount"="0"
"mysql.port"="4000"
"mysql.db"="sink_retry"
"mysql.host"="127.0.0.1"
"dotransactions"="false"
"updateproportion"="0"
"scanproportion"="0"
"recordcount"="10"
"readallfields"="true"
"mysql.user"="root"
"workload"="core"
"requestdistribution"="uniform"
"insertproportion"="0"
**********************************************
Run finished, takes 11.259266ms
INSERT - Takes(s): 0.0, Count: 10, OPS: 1783.9, Avg(us): 2171, Min(us): 1171, Max(us): 5561, 95th(us): 6000, 99th(us): 6000
[Sat May  4 19:43:04 CST 2024] <<<<<< START cdc server in sink_retry case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ (( i = 0 ))
+ (( i <= 50 ))
+ GO_FAILPOINTS='github.com/pingcap/tiflow/cdc/sink/dmlsink/txn/mysql/MySQLSinkTxnRandomError=25%return(true)'
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.sink_retry.28892891.out server --log-file /tmp/tidb_cdc_test/sink_retry/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/sink_retry/cdc_data --cluster-id default
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:43:04 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/d599afa4-b12f-4397-accd-e953df6f0e22
	{"id":"d599afa4-b12f-4397-accd-e953df6f0e22","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822981}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369efa8c5
	d599afa4-b12f-4397-accd-e953df6f0e22

/tidb/cdc/default/default/upstream/7365108570950873885
	{"id":7365108570950873885,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/d599afa4-b12f-4397-accd-e953df6f0e22
	{"id":"d599afa4-b12f-4397-accd-e953df6f0e22","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822981}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369efa8c5
	d599afa4-b12f-4397-accd-e953df6f0e22

/tidb/cdc/default/default/upstream/7365108570950873885
	{"id":7365108570950873885,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/d599afa4-b12f-4397-accd-e953df6f0e22
	{"id":"d599afa4-b12f-4397-accd-e953df6f0e22","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822981}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369efa8c5
	d599afa4-b12f-4397-accd-e953df6f0e22

/tidb/cdc/default/default/upstream/7365108570950873885
	{"id":7365108570950873885,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
check_changefeed_status 127.0.0.1:8300 c5073bcc-4f80-43b5-8b38-d5bcab5984e5 warning last_warning kafka sink injected error
+ endpoint=127.0.0.1:8300
+ changefeed_id=c5073bcc-4f80-43b5-8b38-d5bcab5984e5
+ expected_state=warning
+ field=last_warning
+ error_pattern=kafka
++ curl 127.0.0.1:8300/api/v2/changefeeds/c5073bcc-4f80-43b5-8b38-d5bcab5984e5/status
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100   244  100   244    0     0   2190      0 --:--:-- --:--:-- --:--:--  2178
100   244  100   244    0     0   2188      0 --:--:-- --:--:-- --:--:--  2178
+ info='{"state":"warning","resolved_ts":449530556331851777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}'
+ echo '{"state":"warning","resolved_ts":449530556331851777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}'
{"state":"warning","resolved_ts":449530556331851777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}
++ echo '{"state":"warning","resolved_ts":449530556331851777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka' sink injected 'error"}}'
++ jq -r .state
+ state=warning
+ [[ ! warning == \w\a\r\n\i\n\g ]]
+ [[ -z last_warning ]]
++ echo '{"state":"warning","resolved_ts":449530556331851777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka' sink injected 'error"}}'
++ jq -r .last_warning.message
+ error_msg='kafka sink injected error'
+ [[ ! kafka sink injected error =~ kafka ]]
run task successfully
[Sat May  4 19:43:04 CST 2024] <<<<<< START kafka consumer in gc_safepoint case >>>>>>
0
check_changefeed_status 127.0.0.1:8300 c5073bcc-4f80-43b5-8b38-d5bcab5984e5 normal
+ endpoint=127.0.0.1:8300
+ changefeed_id=c5073bcc-4f80-43b5-8b38-d5bcab5984e5
+ expected_state=normal
+ field=
+ error_pattern=
++ curl 127.0.0.1:8300/api/v2/changefeeds/c5073bcc-4f80-43b5-8b38-d5bcab5984e5/status
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100   244  100   244    0     0   1820      0 --:--:-- --:--:-- --:--:--  1834
+ info='{"state":"warning","resolved_ts":449530556331851777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}'
+ echo '{"state":"warning","resolved_ts":449530556331851777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}'
{"state":"warning","resolved_ts":449530556331851777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}
++ echo '{"state":"warning","resolved_ts":449530556331851777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka' sink injected 'error"}}'
++ jq -r .state
+ state=warning
+ [[ ! warning == \n\o\r\m\a\l ]]
+ echo 'changefeed state warning does not equal to normal'
changefeed state warning does not equal to normal
+ exit 1
run task failed 1-th time, retry later
+ set +x
+ tso='449530556229353474
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530556229353474 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
[Sat May  4 19:43:05 CST 2024] <<<<<< START cdc server in resolve_lock case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ GO_FAILPOINTS=
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.resolve_lock.29292931.out server --log-file /tmp/tidb_cdc_test/resolve_lock/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/resolve_lock/cdc_data --cluster-id default
+ [[ no != \n\o ]]
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
table ddl_manager.finish_mark not exists for 48-th check, retry later
check diff failed 1-th time, retry later
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/consistent_partition_table/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:43:05 CST 2024] <<<<<< run test case consistent_partition_table success! >>>>>>
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)

+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:43:05 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/99da215b-d7af-47b1-b43c-50f7b2d387d6
	{"id":"99da215b-d7af-47b1-b43c-50f7b2d387d6","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822983}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369ee82cf
	99da215b-d7af-47b1-b43c-50f7b2d387d6

/tidb/cdc/default/default/upstream/7365108574218686803
	{"id":7365108574218686803,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/99da215b-d7af-47b1-b43c-50f7b2d387d6
	{"id":"99da215b-d7af-47b1-b43c-50f7b2d387d6","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822983}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369ee82cf
	99da215b-d7af-47b1-b43c-50f7b2d387d6

/tidb/cdc/default/default/upstream/7365108574218686803
+ grep -q 'failed to get info:'
	{"id":7365108574218686803,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/99da215b-d7af-47b1-b43c-50f7b2d387d6
	{"id":"99da215b-d7af-47b1-b43c-50f7b2d387d6","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822983}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369ee82cf
	99da215b-d7af-47b1-b43c-50f7b2d387d6

/tidb/cdc/default/default/upstream/7365108574218686803
	{"id":7365108574218686803,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.default_value.cli.2996.out cli changefeed create --start-ts=449530555575304193 '--sink-uri=kafka://127.0.0.1:9092/ticdc-default-value-test-9256?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760'
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:43:06 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/689478be-48c4-44b6-b087-90997e3976c1
	{"id":"689478be-48c4-44b6-b087-90997e3976c1","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822983}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369e5d265
	689478be-48c4-44b6-b087-90997e3976c1

/tidb/cdc/default/default/upstream/7365108565155202636
	{"id":7365108565155202636,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/689478be-48c4-44b6-b087-90997e3976c1
	{"id":"689478be-48c4-44b6-b087-90997e3976c1","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822983}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369e5d265
	689478be-48c4-44b6-b087-90997e3976c1

/tidb/cdc/default/default/upstream/7365108565155202636
	{"id":7365108565155202636,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/689478be-48c4-44b6-b087-90997e3976c1
	{"id":"689478be-48c4-44b6-b087-90997e3976c1","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822983}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369e5d265
	689478be-48c4-44b6-b087-90997e3976c1

/tidb/cdc/default/default/upstream/7365108565155202636
	{"id":7365108565155202636,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
Create changefeed successfully!
ID: 26493080-7db8-440f-880b-d7b9a09b9482
Info: {"upstream_id":7365108574218686803,"namespace":"default","id":"26493080-7db8-440f-880b-d7b9a09b9482","sink_uri":"kafka://127.0.0.1:9092/ticdc-default-value-test-9256?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:43:06.2232579+08:00","start_ts":449530555575304193,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530555575304193,"checkpoint_ts":449530555575304193,"checkpoint_time":"2024-05-04 19:43:01.168"}
PASS
[Sat May  4 19:43:06 CST 2024] <<<<<< START kafka consumer in processor_err_chan case >>>>>>
check_changefeed_state http://127.0.0.1:2379 70056772-63fa-4e7d-b9ab-b27cc0a3a043 normal null
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=70056772-63fa-4e7d-b9ab-b27cc0a3a043
+ expected_state=normal
+ error_msg=null
+ tls_dir=null
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c 70056772-63fa-4e7d-b9ab-b27cc0a3a043 -s
+ info='{
  "upstream_id": 7365108565155202636,
  "namespace": "default",
  "id": "70056772-63fa-4e7d-b9ab-b27cc0a3a043",
  "state": "normal",
  "checkpoint_tso": 449530556866625541,
  "checkpoint_time": "2024-05-04 19:43:06.094",
  "error": null
}'
+ echo '{
  "upstream_id": 7365108565155202636,
  "namespace": "default",
  "id": "70056772-63fa-4e7d-b9ab-b27cc0a3a043",
  "state": "normal",
  "checkpoint_tso": 449530556866625541,
  "checkpoint_time": "2024-05-04 19:43:06.094",
  "error": null
}'
{
  "upstream_id": 7365108565155202636,
  "namespace": "default",
  "id": "70056772-63fa-4e7d-b9ab-b27cc0a3a043",
  "state": "normal",
  "checkpoint_tso": 449530556866625541,
  "checkpoint_time": "2024-05-04 19:43:06.094",
  "error": null
}
++ echo '{' '"upstream_id":' 7365108565155202636, '"namespace":' '"default",' '"id":' '"70056772-63fa-4e7d-b9ab-b27cc0a3a043",' '"state":' '"normal",' '"checkpoint_tso":' 449530556866625541, '"checkpoint_time":' '"2024-05-04' '19:43:06.094",' '"error":' null '}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \n\o\r\m\a\l ]]
++ echo '{' '"upstream_id":' 7365108565155202636, '"namespace":' '"default",' '"id":' '"70056772-63fa-4e7d-b9ab-b27cc0a3a043",' '"state":' '"normal",' '"checkpoint_tso":' 449530556866625541, '"checkpoint_time":' '"2024-05-04' '19:43:06.094",' '"error":' null '}'
++ jq -r .error.message
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
+ message=null
+ [[ ! null =~ null ]]
run task successfully
check diff failed 1-th time, retry later
check_changefeed_status 127.0.0.1:8300 c5073bcc-4f80-43b5-8b38-d5bcab5984e5 normal
+ endpoint=127.0.0.1:8300
+ changefeed_id=c5073bcc-4f80-43b5-8b38-d5bcab5984e5
+ expected_state=normal
+ field=
+ error_pattern=
++ curl 127.0.0.1:8300/api/v2/changefeeds/c5073bcc-4f80-43b5-8b38-d5bcab5984e5/status
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100   244  100   244    0     0   3983      0 --:--:-- --:--:-- --:--:--  4000
+ info='{"state":"warning","resolved_ts":449530556856139777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}'
+ echo '{"state":"warning","resolved_ts":449530556856139777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}'
{"state":"warning","resolved_ts":449530556856139777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}
++ echo '{"state":"warning","resolved_ts":449530556856139777,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka' sink injected 'error"}}'
++ jq -r .state
+ state=warning
+ [[ ! warning == \n\o\r\m\a\l ]]
+ echo 'changefeed state warning does not equal to normal'
changefeed state warning does not equal to normal
+ exit 1
run task failed 2-th time, retry later
+ set +x
[Sat May  4 19:43:07 CST 2024] <<<<<< START kafka consumer in default_value case >>>>>>
go: downloading github.com/pingcap/errors v0.11.5-0.20240318064555-6bd07397691f
go: downloading github.com/pingcap/log v1.1.1-0.20240314023424-862ccc32f18d
go: downloading go.uber.org/zap v1.27.0
go: downloading github.com/google/uuid v1.6.0
go: downloading github.com/BurntSushi/toml v1.3.2
go: downloading github.com/pingcap/tidb-tools v0.0.0-20240305021104-9f9bea84490b
go: downloading github.com/pingcap/tidb v1.1.0-beta.0.20240415145106-cd9c676e9ba4
go: downloading github.com/pingcap/failpoint v0.0.0-20220801062533-2eaa32854a6c
go: downloading golang.org/x/time v0.5.0
go: downloading golang.org/x/sync v0.7.0
go: downloading google.golang.org/grpc v1.62.1
go: downloading github.com/go-sql-driver/mysql v1.7.1
go: downloading gopkg.in/natefinch/lumberjack.v2 v2.2.1
go: downloading go.uber.org/atomic v1.11.0
go: downloading go.uber.org/multierr v1.11.0
go: downloading github.com/pingcap/tidb/pkg/parser v0.0.0-20240410110152-5fc42c9be2f5
go: downloading github.com/coreos/go-semver v0.3.1
table ddl_manager.finish_mark not exists for 49-th check, retry later
check diff failed 2-th time, retry later
go: downloading google.golang.org/genproto/googleapis/rpc v0.0.0-20240401170217-c3f982113cda
go: downloading google.golang.org/protobuf v1.33.0
go: downloading golang.org/x/sys v0.19.0
go: downloading golang.org/x/net v0.24.0
go: downloading github.com/golang/protobuf v1.5.4
go: downloading google.golang.org/genproto v0.0.0-20240401170217-c3f982113cda
go: downloading golang.org/x/text v0.14.0
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:43:08 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/ef12fc43-5190-4cca-801e-d58d400669d6
	{"id":"ef12fc43-5190-4cca-801e-d58d400669d6","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822985}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369f61acf
	ef12fc43-5190-4cca-801e-d58d400669d6

/tidb/cdc/default/default/upstream/7365108576170879301
	{"id":7365108576170879301,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/ef12fc43-5190-4cca-801e-d58d400669d6
	{"id":"ef12fc43-5190-4cca-801e-d58d400669d6","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822985}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369f61acf
	ef12fc43-5190-4cca-801e-d58d400669d6

/tidb/cdc/default/default/upstream/7365108576170879301
	{"id":7365108576170879301,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/ef12fc43-5190-4cca-801e-d58d400669d6
	{"id":"ef12fc43-5190-4cca-801e-d58d400669d6","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822985}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369f61acf
	ef12fc43-5190-4cca-801e-d58d400669d6

/tidb/cdc/default/default/upstream/7365108576170879301
	{"id":7365108576170879301,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.resolve_lock.cli.2988.out cli changefeed create --start-ts=449530556229353474 '--sink-uri=kafka://127.0.0.1:9092/ticdc-resolve-lock-test-10490?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760'
Create changefeed successfully!
ID: b1355e00-fa0e-44bc-b5ec-77c0274a21f8
Info: {"upstream_id":7365108576170879301,"namespace":"default","id":"b1355e00-fa0e-44bc-b5ec-77c0274a21f8","sink_uri":"kafka://127.0.0.1:9092/ticdc-resolve-lock-test-10490?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:43:08.708727897+08:00","start_ts":449530556229353474,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530556229353474,"checkpoint_ts":449530556229353474,"checkpoint_time":"2024-05-04 19:43:03.663"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/kafka_big_messages_v2/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:43:08 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/fc0f3e7f-7b76-40f1-8279-014b88269102
	{"id":"fc0f3e7f-7b76-40f1-8279-014b88269102","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822984}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369e9ccf5
	fc0f3e7f-7b76-40f1-8279-014b88269102

/tidb/cdc/default/default/upstream/7365108561787137439
	{"id":7365108561787137439,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/fc0f3e7f-7b76-40f1-8279-014b88269102
	{"id":"fc0f3e7f-7b76-40f1-8279-014b88269102","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822984}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369e9ccf5
	fc0f3e7f-7b76-40f1-8279-014b88269102

/tidb/cdc/default/default/upstream/7365108561787137439
	{"id":7365108561787137439,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/fc0f3e7f-7b76-40f1-8279-014b88269102
	{"id":"fc0f3e7f-7b76-40f1-8279-014b88269102","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714822984}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f4369e9ccf5
	fc0f3e7f-7b76-40f1-8279-014b88269102

/tidb/cdc/default/default/upstream/7365108561787137439
	{"id":7365108561787137439,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.sink_retry.cli.2946.out cli changefeed create --start-ts=449530555652112385 '--sink-uri=kafka://127.0.0.1:9092/ticdc-sink-retry-test-2783?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760'
Create changefeed successfully!
ID: 5512dc99-9f41-43a9-81ed-a0d23dbbd20c
Info: {"upstream_id":7365108561787137439,"namespace":"default","id":"5512dc99-9f41-43a9-81ed-a0d23dbbd20c","sink_uri":"kafka://127.0.0.1:9092/ticdc-sink-retry-test-2783?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:43:09.070567659+08:00","start_ts":449530555652112385,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530555652112385,"checkpoint_ts":449530555652112385,"checkpoint_time":"2024-05-04 19:43:01.461"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
check diff failed 2-th time, retry later
table ddl_manager.finish_mark not exists for 50-th check, retry later
+ set +x
[Sat May  4 19:43:10 CST 2024] <<<<<< START kafka consumer in resolve_lock case >>>>>>
go: downloading github.com/pingcap/errors v0.11.5-0.20240318064555-6bd07397691f
go: downloading github.com/pingcap/tidb/pkg/parser v0.0.0-20240410110152-5fc42c9be2f5
go: downloading github.com/pingcap/tidb v1.1.0-beta.0.20240415145106-cd9c676e9ba4
go: downloading github.com/pingcap/kvproto v0.0.0-20240227073058-929ab83f9754
go: downloading github.com/tikv/client-go/v2 v2.0.8-0.20240409022718-714958ccd4d5
go: downloading github.com/tikv/pd/client v0.0.0-20240322051414-fb9e2d561b6e
go: downloading github.com/pingcap/log v1.1.1-0.20240314023424-862ccc32f18d
go: downloading github.com/pingcap/tidb-tools v0.0.0-20240305021104-9f9bea84490b
go: downloading go.uber.org/zap v1.27.0
go: downloading github.com/BurntSushi/toml v1.3.2
go: downloading go.uber.org/atomic v1.11.0
go: downloading gopkg.in/natefinch/lumberjack.v2 v2.2.1
go: downloading github.com/cznic/mathutil v0.0.0-20181122101859-297441e03548
go: downloading golang.org/x/exp v0.0.0-20240409090435-93d18d7e34b8
go: downloading go.etcd.io/etcd/client/v3 v3.5.12
go: downloading github.com/pingcap/tipb v0.0.0-20240318032315-55a7867ddd50
go: downloading github.com/influxdata/tdigest v0.0.1
go: downloading github.com/prometheus/client_golang v1.19.0
go: downloading github.com/spf13/pflag v1.0.5
go: downloading github.com/pingcap/kvproto v0.0.0-20240227073058-929ab83f9754
go: downloading github.com/danjacques/gofslock v0.0.0-20240212154529-d899e02bfe22
go: downloading github.com/tikv/pd/client v0.0.0-20240322051414-fb9e2d561b6e
go: downloading github.com/opentracing/opentracing-go v1.2.0
go: downloading github.com/grpc-ecosystem/go-grpc-middleware v1.4.0
go: downloading github.com/uber/jaeger-client-go v2.30.0+incompatible
go: downloading gopkg.in/yaml.v2 v2.4.0
go: downloading github.com/tikv/client-go/v2 v2.0.8-0.20240409022718-714958ccd4d5
go: downloading github.com/ngaut/pools v0.0.0-20180318154953-b7bc8c42aac7
go: downloading github.com/docker/go-units v0.5.0
go: downloading github.com/tiancaiamao/gp v0.0.0-20221230034425-4025bc8a4d4a
go: downloading github.com/prometheus/client_model v0.6.1
go: downloading github.com/jellydator/ttlcache/v3 v3.0.1
go: downloading github.com/coocood/freecache v1.2.1
go: downloading github.com/pingcap/sysutil v1.0.1-0.20240311050922-ae81ee01f3a5
go: downloading github.com/stretchr/testify v1.9.0
go: downloading github.com/shirou/gopsutil/v3 v3.24.2
go: downloading github.com/cockroachdb/errors v1.11.1
go: downloading cloud.google.com/go/storage v1.39.1
go: downloading github.com/Azure/azure-sdk-for-go/sdk/azcore v1.9.1
go: downloading github.com/scalalang2/golang-fifo v0.1.5
go: downloading github.com/Azure/azure-sdk-for-go/sdk/azidentity v1.5.1
go: downloading github.com/Azure/azure-sdk-for-go/sdk/storage/azblob v1.0.0
go: downloading github.com/twmb/murmur3 v1.1.6
go: downloading github.com/gorilla/mux v1.8.0
go: downloading github.com/tidwall/btree v1.7.0
go: downloading github.com/aliyun/alibaba-cloud-sdk-go v1.61.1581
go: downloading github.com/aws/aws-sdk-go v1.50.0
go: downloading github.com/opentracing/basictracer-go v1.1.0
go: downloading github.com/tikv/pd v1.1.0-beta.0.20240407022249-7179657d129b
go: downloading github.com/go-resty/resty/v2 v2.11.0
go: downloading github.com/klauspost/compress v1.17.8
go: downloading go.etcd.io/etcd/api/v3 v3.5.12
go: downloading github.com/ks3sdklib/aws-sdk-go v1.2.9
go: downloading github.com/yangkeao/ldap/v3 v3.4.5-0.20230421065457-369a3bab1117
go: downloading cloud.google.com/go v0.112.2
go: downloading golang.org/x/oauth2 v0.18.0
go: downloading google.golang.org/api v0.170.0
go: downloading go.uber.org/mock v0.4.0
go: downloading github.com/google/btree v1.1.2
go: downloading github.com/gogo/protobuf v1.3.2
go: downloading golang.org/x/tools v0.20.0
go: downloading github.com/cockroachdb/pebble v1.1.0
go: downloading github.com/jfcg/sorty/v2 v2.1.0
go: downloading github.com/ngaut/sync2 v0.0.0-20141008032647-7a24ed77b2ef
go: downloading github.com/cespare/xxhash/v2 v2.3.0
go: downloading github.com/joho/sqltocsv v0.0.0-20210428211105-a6d6801d59df
go: downloading github.com/carlmjohnson/flagext v0.21.0
go: downloading github.com/jedib0t/go-pretty/v6 v6.2.2
go: downloading github.com/dolthub/swiss v0.2.1
go: downloading github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec
go: downloading github.com/dgraph-io/ristretto v0.1.1
go: downloading github.com/golang/snappy v0.0.4
go: downloading github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc
go: downloading gopkg.in/yaml.v3 v3.0.1
go: downloading github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2
go: downloading github.com/Azure/azure-sdk-for-go/sdk/internal v1.5.1
go: downloading github.com/AzureAD/microsoft-authentication-library-for-go v1.2.1
go: downloading golang.org/x/crypto v0.22.0
go: downloading go.etcd.io/etcd/client/pkg/v3 v3.5.12
go: downloading github.com/beorn7/perks v1.0.1
go: downloading github.com/prometheus/common v0.52.2
go: downloading github.com/prometheus/procfs v0.13.0
go: downloading github.com/pkg/errors v0.9.1
go: downloading github.com/uber/jaeger-lib v2.4.1+incompatible
go: downloading github.com/Azure/go-ntlmssp v0.0.0-20221128193559-754e69321358
go: downloading github.com/go-asn1-ber/asn1-ber v1.5.4
go: downloading github.com/cockroachdb/logtags v0.0.0-20230118201751-21c54148d20b
go: downloading github.com/cockroachdb/redact v1.1.5
go: downloading github.com/getsentry/sentry-go v0.27.0
go: downloading github.com/cloudfoundry/gosigar v1.3.6
go: downloading github.com/lestrrat-go/jwx/v2 v2.0.21
go: downloading github.com/dgryski/go-farm v0.0.0-20200201041132-a6ae2369ad13
go: downloading github.com/spkg/bom v1.0.0
go: downloading github.com/xitongsys/parquet-go v1.6.0
go: downloading github.com/otiai10/copy v1.2.0
go: downloading github.com/asaskevich/govalidator v0.0.0-20230301143203-a9d515a09cc2
go: downloading github.com/tklauser/go-sysconf v0.3.12
go: downloading github.com/jfcg/sixb v1.3.8
go: downloading google.golang.org/genproto/googleapis/api v0.0.0-20240401170217-c3f982113cda
go: downloading github.com/google/pprof v0.0.0-20240117000934-35fc243c5815
go: downloading github.com/wangjohn/quickselect v0.0.0-20161129230411-ed8402a42d5f
go: downloading github.com/dolthub/maphash v0.1.0
go: downloading cloud.google.com/go/compute/metadata v0.2.3
go: downloading cloud.google.com/go/iam v1.1.7
go: downloading github.com/googleapis/gax-go/v2 v2.12.3
go: downloading cloud.google.com/go/compute v1.25.1
go: downloading github.com/kr/pretty v0.3.1
go: downloading github.com/cheggaaa/pb/v3 v3.0.8
go: downloading github.com/robfig/cron/v3 v3.0.1
go: downloading github.com/coreos/go-systemd/v22 v22.5.0
go: downloading github.com/pingcap/goleveldb v0.0.0-20191226122134-f82aafb29989
go: downloading github.com/robfig/cron v1.2.0
go: downloading github.com/pingcap/badger v1.5.1-0.20230103063557-828f39b09b6d
go: downloading github.com/mattn/go-runewidth v0.0.15
go: downloading github.com/kylelemons/godebug v1.1.0
go: downloading github.com/pkg/browser v0.0.0-20240102092130-5ac0b6a4141c
go: downloading go.opencensus.io v0.23.1-0.20220331163232-052120675fac
go: downloading go.opentelemetry.io/otel v1.24.0
go: downloading go.opentelemetry.io/otel/trace v1.24.0
go: downloading github.com/tklauser/numcpus v0.6.1
go: downloading github.com/apache/thrift v0.16.0
go: downloading github.com/kr/text v0.2.0
go: downloading github.com/rogpeppe/go-internal v1.12.0
go: downloading github.com/VividCortex/ewma v1.2.0
go: downloading github.com/tikv/pd v1.1.0-beta.0.20240407022249-7179657d129b
go: downloading go.uber.org/multierr v1.11.0
go: downloading github.com/pingcap/failpoint v0.0.0-20220801062533-2eaa32854a6c
go: downloading github.com/coreos/go-semver v0.3.1
go: downloading github.com/go-sql-driver/mysql v1.7.1
go: downloading google.golang.org/grpc v1.62.1
go: downloading github.com/pkg/errors v0.9.1
go: downloading github.com/opentracing/opentracing-go v1.2.0
go: downloading github.com/google/uuid v1.6.0
go: downloading github.com/tiancaiamao/gp v0.0.0-20221230034425-4025bc8a4d4a
go: downloading github.com/twmb/murmur3 v1.1.6
go: downloading github.com/cznic/mathutil v0.0.0-20181122101859-297441e03548
go: downloading go.etcd.io/etcd/api/v3 v3.5.12
go: downloading go.etcd.io/etcd/client/v3 v3.5.12
go: downloading github.com/prometheus/client_golang v1.19.0
go: downloading github.com/prometheus/client_model v0.6.1
go: downloading github.com/golang/protobuf v1.5.4
go: downloading github.com/dgryski/go-farm v0.0.0-20200201041132-a6ae2369ad13
go: downloading github.com/docker/go-units v0.5.0
go: downloading github.com/gogo/protobuf v1.3.2
go: downloading github.com/google/btree v1.1.2
go: downloading github.com/grpc-ecosystem/go-grpc-middleware v1.4.0
go: downloading golang.org/x/sync v0.7.0
go: downloading google.golang.org/protobuf v1.33.0
go: downloading github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec
go: downloading go.etcd.io/etcd/client/pkg/v3 v3.5.12
go: downloading google.golang.org/genproto/googleapis/api v0.0.0-20240401170217-c3f982113cda
go: downloading google.golang.org/genproto v0.0.0-20240401170217-c3f982113cda
go: downloading github.com/fatih/color v1.16.0
go: downloading github.com/mattn/go-colorable v0.1.13
go: downloading github.com/mattn/go-isatty v0.0.20
go: downloading github.com/lestrrat-go/blackmagic v1.0.2
go: downloading github.com/lestrrat-go/httprc v1.0.5
go: downloading github.com/lestrrat-go/iter v1.0.2
go: downloading github.com/lestrrat-go/option v1.0.1
go: downloading github.com/golang-jwt/jwt/v5 v5.2.0
go: downloading github.com/dustin/go-humanize v1.0.1
go: downloading github.com/golang/glog v1.2.0
go: downloading github.com/rivo/uniseg v0.4.7
go: downloading github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da
go: downloading github.com/lestrrat-go/httpcc v1.0.1
go: downloading github.com/ncw/directio v1.0.5
go: downloading github.com/coocood/rtutil v0.0.0-20190304133409-c84515f646f2
go: downloading github.com/coocood/bbloom v0.0.0-20190830030839-58deb6228d64
go: downloading github.com/klauspost/cpuid v1.3.1
go: downloading github.com/golang-jwt/jwt v3.2.2+incompatible
go: downloading github.com/go-logr/logr v1.4.1
go: downloading go.opentelemetry.io/otel/metric v1.24.0
go: downloading github.com/go-logr/stdr v1.2.2
go: downloading github.com/coreos/go-systemd/v22 v22.5.0
go: downloading github.com/beorn7/perks v1.0.1
go: downloading github.com/cespare/xxhash/v2 v2.3.0
go: downloading github.com/prometheus/common v0.52.2
go: downloading github.com/prometheus/procfs v0.13.0
go: downloading github.com/cloudfoundry/gosigar v1.3.6
go: downloading golang.org/x/exp v0.0.0-20240409090435-93d18d7e34b8
go: downloading golang.org/x/sys v0.19.0
go: downloading golang.org/x/net v0.24.0
go: downloading google.golang.org/genproto/googleapis/rpc v0.0.0-20240401170217-c3f982113cda
+ set +x
[Sat May  4 19:43:10 CST 2024] <<<<<< START kafka consumer in sink_retry case >>>>>>
go: downloading github.com/DataDog/zstd v1.5.5
go: downloading github.com/cockroachdb/tokenbucket v0.0.0-20230807174530-cc333fc44b06
go: downloading golang.org/x/text v0.14.0
check diff failed 3-th time, retry later
check_changefeed_status 127.0.0.1:8300 c5073bcc-4f80-43b5-8b38-d5bcab5984e5 normal
+ endpoint=127.0.0.1:8300
+ changefeed_id=c5073bcc-4f80-43b5-8b38-d5bcab5984e5
+ expected_state=normal
+ field=
+ error_pattern=
++ curl 127.0.0.1:8300/api/v2/changefeeds/c5073bcc-4f80-43b5-8b38-d5bcab5984e5/status
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100   244  100   244    0     0   1536      0 --:--:-- --:--:-- --:--:--  1534
100   244  100   244    0     0   1535      0 --:--:-- --:--:-- --:--:--  1534
+ info='{"state":"warning","resolved_ts":449530558166859778,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}'
+ echo '{"state":"warning","resolved_ts":449530558166859778,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}'
{"state":"warning","resolved_ts":449530558166859778,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka sink injected error"}}
++ echo '{"state":"warning","resolved_ts":449530558166859778,"checkpoint_ts":449530554693451804,"last_warning":{"time":"2024-05-04T19:43:01.708422741+08:00","addr":"127.0.0.1:8300","code":"CDC:ErrProcessorUnknown","message":"kafka' sink injected 'error"}}'
++ jq -r .state
+ state=warning
+ [[ ! warning == \n\o\r\m\a\l ]]
+ echo 'changefeed state warning does not equal to normal'
changefeed state warning does not equal to normal
+ exit 1
run task failed 3-th time, retry later
check diff successfully
check_safepoint_forward http://127.0.0.1:2379 7365108570950873885 449530557854121988 449530556386377732
run task successfully
check_changefeed_state http://127.0.0.1:2379 e08dd9ed-4bf7-4060-8e52-d1d52c1d6226 stopped null
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=e08dd9ed-4bf7-4060-8e52-d1d52c1d6226
+ expected_state=stopped
+ error_msg=null
+ tls_dir=null
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c e08dd9ed-4bf7-4060-8e52-d1d52c1d6226 -s
go: downloading github.com/opentracing/basictracer-go v1.1.0
go: downloading github.com/pingcap/tipb v0.0.0-20240318032315-55a7867ddd50
go: downloading github.com/coocood/freecache v1.2.1
go: downloading github.com/dgraph-io/ristretto v0.1.1
go: downloading gopkg.in/yaml.v2 v2.4.0
go: downloading github.com/spf13/pflag v1.0.5
go: downloading github.com/influxdata/tdigest v0.0.1
go: downloading github.com/pingcap/sysutil v1.0.1-0.20240311050922-ae81ee01f3a5
go: downloading github.com/ngaut/pools v0.0.0-20180318154953-b7bc8c42aac7
go: downloading github.com/uber/jaeger-client-go v2.30.0+incompatible
go: downloading github.com/danjacques/gofslock v0.0.0-20240212154529-d899e02bfe22
go: downloading github.com/shirou/gopsutil/v3 v3.24.2
go: downloading github.com/gorilla/mux v1.8.0
go: downloading github.com/jellydator/ttlcache/v3 v3.0.1
go: downloading github.com/cockroachdb/errors v1.11.1
go: downloading github.com/dolthub/swiss v0.2.1
go: downloading github.com/yangkeao/ldap/v3 v3.4.5-0.20230421065457-369a3bab1117
go: downloading cloud.google.com/go/storage v1.39.1
go: downloading github.com/stretchr/testify v1.9.0
go: downloading github.com/Azure/azure-sdk-for-go/sdk/azcore v1.9.1
go: downloading github.com/otiai10/copy v1.2.0
go: downloading github.com/scalalang2/golang-fifo v0.1.5
go: downloading go.uber.org/mock v0.4.0
go: downloading github.com/Azure/azure-sdk-for-go/sdk/azidentity v1.5.1
go: downloading github.com/Azure/azure-sdk-for-go/sdk/storage/azblob v1.0.0
go: downloading github.com/tidwall/btree v1.7.0
go: downloading github.com/aliyun/alibaba-cloud-sdk-go v1.61.1581
go: downloading github.com/aws/aws-sdk-go v1.50.0
go: downloading github.com/golang/snappy v0.0.4
go: downloading github.com/asaskevich/govalidator v0.0.0-20230301143203-a9d515a09cc2
go: downloading github.com/go-resty/resty/v2 v2.11.0
go: downloading github.com/klauspost/compress v1.17.8
go: downloading github.com/cockroachdb/pebble v1.1.0
go: downloading golang.org/x/tools v0.20.0
go: downloading github.com/jfcg/sorty/v2 v2.1.0
go: downloading github.com/ks3sdklib/aws-sdk-go v1.2.9
go: downloading golang.org/x/oauth2 v0.18.0
go: downloading google.golang.org/api v0.170.0
go: downloading golang.org/x/time v0.5.0
go: downloading github.com/jedib0t/go-pretty/v6 v6.2.2
go: downloading github.com/carlmjohnson/flagext v0.21.0
go: downloading github.com/cheggaaa/pb/v3 v3.0.8
go: downloading github.com/spkg/bom v1.0.0
go: downloading github.com/xitongsys/parquet-go v1.6.0
go: downloading github.com/pingcap/badger v1.5.1-0.20230103063557-828f39b09b6d
go: downloading github.com/joho/sqltocsv v0.0.0-20210428211105-a6d6801d59df
go: downloading github.com/sourcegraph/appdash v0.0.0-20190731080439-ebfcffb1b5c0
go: downloading github.com/ngaut/sync2 v0.0.0-20141008032647-7a24ed77b2ef
go: downloading github.com/lestrrat-go/jwx/v2 v2.0.21
go: downloading github.com/dolthub/maphash v0.1.0
go: downloading github.com/pingcap/goleveldb v0.0.0-20191226122134-f82aafb29989
go: downloading github.com/Azure/go-ntlmssp v0.0.0-20221128193559-754e69321358
go: downloading github.com/jfcg/sixb v1.3.8
go: downloading github.com/go-asn1-ber/asn1-ber v1.5.4
go: downloading github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc
go: downloading github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2
go: downloading gopkg.in/yaml.v3 v3.0.1
go: downloading github.com/fatih/color v1.16.0
go: downloading github.com/vbauerster/mpb/v7 v7.5.3
go: downloading cloud.google.com/go v0.112.2
go: downloading golang.org/x/term v0.19.0
go: downloading github.com/spf13/cobra v1.8.0
go: downloading github.com/Azure/azure-sdk-for-go/sdk/internal v1.5.1
go: downloading github.com/AzureAD/microsoft-authentication-library-for-go v1.2.1
go: downloading golang.org/x/crypto v0.22.0
go: downloading github.com/VividCortex/ewma v1.2.0
go: downloading github.com/mattn/go-runewidth v0.0.15
go: downloading github.com/mattn/go-isatty v0.0.20
go: downloading github.com/mattn/go-colorable v0.1.13
go: downloading github.com/google/pprof v0.0.0-20240117000934-35fc243c5815
go: downloading github.com/json-iterator/go v1.1.12
go: downloading cloud.google.com/go/compute/metadata v0.2.3
go: downloading github.com/apache/thrift v0.16.0
go: downloading github.com/robfig/cron/v3 v3.0.1
go: downloading github.com/wangjohn/quickselect v0.0.0-20161129230411-ed8402a42d5f
go: downloading github.com/uber/jaeger-lib v2.4.1+incompatible
go: downloading github.com/Masterminds/semver v1.5.0
go: downloading cloud.google.com/go/compute v1.25.1
go: downloading k8s.io/api v0.28.6
go: downloading github.com/robfig/cron v1.2.0
go: downloading github.com/emirpasic/gods v1.18.1
go: downloading github.com/cockroachdb/logtags v0.0.0-20230118201751-21c54148d20b
go: downloading github.com/cockroachdb/redact v1.1.5
go: downloading github.com/getsentry/sentry-go v0.27.0
go: downloading github.com/tklauser/go-sysconf v0.3.12
go: downloading github.com/dustin/go-humanize v1.0.1
go: downloading github.com/golang/glog v1.2.0
go: downloading github.com/acarl005/stripansi v0.0.0-20180116102854-5a71ef0e047d
go: downloading github.com/rivo/uniseg v0.4.7
go: downloading github.com/kr/pretty v0.3.1
go: downloading github.com/kylelemons/godebug v1.1.0
go: downloading github.com/pkg/browser v0.0.0-20240102092130-5ac0b6a4141c
go: downloading github.com/ncw/directio v1.0.5
go: downloading github.com/coocood/rtutil v0.0.0-20190304133409-c84515f646f2
go: downloading github.com/coocood/bbloom v0.0.0-20190830030839-58deb6228d64
go: downloading github.com/golang-jwt/jwt/v5 v5.2.0
go: downloading github.com/modern-go/reflect2 v1.0.2
go: downloading github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd
go: downloading github.com/klauspost/cpuid v1.3.1
go: downloading github.com/kr/text v0.2.0
go: downloading github.com/rogpeppe/go-internal v1.12.0
go: downloading cloud.google.com/go/iam v1.1.7
go: downloading github.com/googleapis/gax-go/v2 v2.12.3
go: downloading go.opencensus.io v0.23.1-0.20220331163232-052120675fac
go: downloading go.opentelemetry.io/otel v1.24.0
go: downloading go.opentelemetry.io/otel/trace v1.24.0
go: downloading github.com/tklauser/numcpus v0.6.1
go: downloading github.com/golang-jwt/jwt v3.2.2+incompatible
go: downloading github.com/lestrrat-go/blackmagic v1.0.2
go: downloading github.com/lestrrat-go/httprc v1.0.5
go: downloading github.com/lestrrat-go/iter v1.0.2
go: downloading github.com/lestrrat-go/option v1.0.1
go: downloading github.com/lestrrat-go/httpcc v1.0.1
+ info='{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",
  "state": "stopped",
  "checkpoint_tso": 449530558378672132,
  "checkpoint_time": "2024-05-04 19:43:11.862",
  "error": null
}'
+ echo '{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",
  "state": "stopped",
  "checkpoint_tso": 449530558378672132,
  "checkpoint_time": "2024-05-04 19:43:11.862",
  "error": null
}'
{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",
  "state": "stopped",
  "checkpoint_tso": 449530558378672132,
  "checkpoint_time": "2024-05-04 19:43:11.862",
  "error": null
}
++ echo '{' '"upstream_id":' 7365108570950873885, '"namespace":' '"default",' '"id":' '"e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",' '"state":' '"stopped",' '"checkpoint_tso":' 449530558378672132, '"checkpoint_time":' '"2024-05-04' '19:43:11.862",' '"error":' null '}'
++ jq -r .state
+ state=stopped
+ [[ ! stopped == \s\t\o\p\p\e\d ]]
++ echo '{' '"upstream_id":' 7365108570950873885, '"namespace":' '"default",' '"id":' '"e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",' '"state":' '"stopped",' '"checkpoint_tso":' 449530558378672132, '"checkpoint_time":' '"2024-05-04' '19:43:11.862",' '"error":' null '}'
++ jq -r .error.message
+ message=null
+ [[ ! null =~ null ]]
run task successfully
check_safepoint_equal http://127.0.0.1:2379 7365108570950873885
table ddl_manager.finish_mark not exists for 51-th check, retry later
go: downloading github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da
go: downloading go.opentelemetry.io/otel/metric v1.24.0
go: downloading github.com/go-logr/logr v1.4.1
go: downloading github.com/go-logr/stdr v1.2.2
check diff successfully
start tidb cluster in /tmp/tidb_cdc_test/kafka_big_messages_v2
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
go: downloading github.com/DataDog/zstd v1.5.5
go: downloading github.com/cockroachdb/tokenbucket v0.0.0-20230807174530-cc333fc44b06
go: downloading k8s.io/apimachinery v0.28.6
go: downloading gopkg.in/inf.v0 v0.9.1
go: downloading sigs.k8s.io/structured-merge-diff/v4 v4.4.1
go: downloading k8s.io/klog/v2 v2.120.1
go: downloading github.com/google/gofuzz v1.2.0
go: downloading k8s.io/utils v0.0.0-20230726121419-3b25d923346b
go: downloading sigs.k8s.io/json v0.0.0-20221116044647-bc3834ca7abd
wait process cdc.test exit for 1-th time...
go: downloading github.com/jmespath/go-jmespath v0.4.0
go: downloading github.com/googleapis/enterprise-certificate-proxy v0.3.2
go: downloading go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp v0.49.0
go: downloading github.com/google/s2a-go v0.1.7
go: downloading go.opentelemetry.io/contrib/instrumentation/google.golang.org/grpc/otelgrpc v0.49.0
go: downloading github.com/felixge/httpsnoop v1.0.4
wait process cdc.test exit for 2-th time...
go: downloading github.com/modern-go/reflect2 v1.0.2
go: downloading github.com/json-iterator/go v1.1.12
go: downloading github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd
table ddl_manager.finish_mark not exists for 52-th check, retry later
cdc.test: no process found
wait process cdc.test exit for 3-th time...
process cdc.test already exit
[Sat May  4 19:43:14 CST 2024] <<<<<< run test case processor_err_chan success! >>>>>>
run task successfully
check_changefeed_state http://127.0.0.1:2379 e08dd9ed-4bf7-4060-8e52-d1d52c1d6226 normal null
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=e08dd9ed-4bf7-4060-8e52-d1d52c1d6226
+ expected_state=normal
+ error_msg=null
+ tls_dir=null
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c e08dd9ed-4bf7-4060-8e52-d1d52c1d6226 -s
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
+ info='{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",
  "state": "normal",
  "checkpoint_tso": 449530558378672132,
  "checkpoint_time": "2024-05-04 19:43:11.862",
  "error": null
}'
+ echo '{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",
  "state": "normal",
  "checkpoint_tso": 449530558378672132,
  "checkpoint_time": "2024-05-04 19:43:11.862",
  "error": null
}'
{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",
  "state": "normal",
  "checkpoint_tso": 449530558378672132,
  "checkpoint_time": "2024-05-04 19:43:11.862",
  "error": null
}
++ echo '{' '"upstream_id":' 7365108570950873885, '"namespace":' '"default",' '"id":' '"e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",' '"state":' '"normal",' '"checkpoint_tso":' 449530558378672132, '"checkpoint_time":' '"2024-05-04' '19:43:11.862",' '"error":' null '}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \n\o\r\m\a\l ]]
++ echo '{' '"upstream_id":' 7365108570950873885, '"namespace":' '"default",' '"id":' '"e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",' '"state":' '"normal",' '"checkpoint_tso":' 449530558378672132, '"checkpoint_time":' '"2024-05-04' '19:43:11.862",' '"error":' null '}'
++ jq -r .error.message
+ message=null
+ [[ ! null =~ null ]]
run task successfully
check_safepoint_forward http://127.0.0.1:2379 7365108570950873885 449530558378672131 449530558378672132
table ddl_manager.finish_mark not exists for 53-th check, retry later
go: downloading github.com/google/s2a-go v0.1.7
go: downloading go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp v0.49.0
go: downloading github.com/googleapis/enterprise-certificate-proxy v0.3.2
go: downloading go.opentelemetry.io/contrib/instrumentation/google.golang.org/grpc/otelgrpc v0.49.0
go: downloading github.com/felixge/httpsnoop v1.0.4
go: downloading github.com/jmespath/go-jmespath v0.4.0
run task successfully
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 13.11 secs (284269815 bytes/sec)
[Pipeline] {
[Pipeline] cache
check_changefeed_state http://127.0.0.1:2379 e08dd9ed-4bf7-4060-8e52-d1d52c1d6226 stopped null
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=e08dd9ed-4bf7-4060-8e52-d1d52c1d6226
+ expected_state=stopped
+ error_msg=null
+ tls_dir=null
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c e08dd9ed-4bf7-4060-8e52-d1d52c1d6226 -s
+ info='{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",
  "state": "stopped",
  "checkpoint_tso": 449530558378672132,
  "checkpoint_time": "2024-05-04 19:43:11.862",
  "error": null
}'
+ echo '{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",
  "state": "stopped",
  "checkpoint_tso": 449530558378672132,
  "checkpoint_time": "2024-05-04 19:43:11.862",
  "error": null
}'
{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",
  "state": "stopped",
  "checkpoint_tso": 449530558378672132,
  "checkpoint_time": "2024-05-04 19:43:11.862",
  "error": null
}
++ echo '{' '"upstream_id":' 7365108570950873885, '"namespace":' '"default",' '"id":' '"e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",' '"state":' '"stopped",' '"checkpoint_tso":' 449530558378672132, '"checkpoint_time":' '"2024-05-04' '19:43:11.862",' '"error":' null '}'
++ jq -r .state
+ state=stopped
+ [[ ! stopped == \s\t\o\p\p\e\d ]]
++ echo '{' '"upstream_id":' 7365108570950873885, '"namespace":' '"default",' '"id":' '"e08dd9ed-4bf7-4060-8e52-d1d52c1d6226",' '"state":' '"stopped",' '"checkpoint_tso":' 449530558378672132, '"checkpoint_time":' '"2024-05-04' '19:43:11.862",' '"error":' null '}'
++ jq -r .error.message
+ message=null
+ [[ ! null =~ null ]]
run task successfully
check_changefeed_status 127.0.0.1:8300 c5073bcc-4f80-43b5-8b38-d5bcab5984e5 normal
+ endpoint=127.0.0.1:8300
+ changefeed_id=c5073bcc-4f80-43b5-8b38-d5bcab5984e5
+ expected_state=normal
+ field=
+ error_pattern=
++ curl 127.0.0.1:8300/api/v2/changefeeds/c5073bcc-4f80-43b5-8b38-d5bcab5984e5/status
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    86  100    86    0     0   1343      0 --:--:-- --:--:-- --:--:--  1365
+ info='{"state":"normal","resolved_ts":449530559739723780,"checkpoint_ts":449530559477579780}'
+ echo '{"state":"normal","resolved_ts":449530559739723780,"checkpoint_ts":449530559477579780}'
{"state":"normal","resolved_ts":449530559739723780,"checkpoint_ts":449530559477579780}
++ echo '{"state":"normal","resolved_ts":449530559739723780,"checkpoint_ts":449530559477579780}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \n\o\r\m\a\l ]]
+ [[ -z '' ]]
++ echo '{"state":"normal","resolved_ts":449530559739723780,"checkpoint_ts":449530559477579780}'
++ jq -r .last_error
+ error_msg=null
+ [[ ! null == \n\u\l\l ]]
++ echo '{"state":"normal","resolved_ts":449530559739723780,"checkpoint_ts":449530559477579780}'
++ jq -r .last_warning
+ error_msg=null
+ [[ ! null == \n\u\l\l ]]
+ exit 0
run task successfully
table kafka_sink_error_resume.t1 exists
table kafka_sink_error_resume.t2 exists
check diff successfully
check diff failed 1-th time, retry later
check_changefeed_state http://127.0.0.1:2379 aa4d62d1-e9c2-4bbb-b784-a2e666a98a39 normal null
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=aa4d62d1-e9c2-4bbb-b784-a2e666a98a39
+ expected_state=normal
+ error_msg=null
+ tls_dir=null
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c aa4d62d1-e9c2-4bbb-b784-a2e666a98a39 -s
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table ddl_manager.finish_mark not exists for 54-th check, retry later
+ info='{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "aa4d62d1-e9c2-4bbb-b784-a2e666a98a39",
  "state": "normal",
  "checkpoint_tso": 449530560069238786,
  "checkpoint_time": "2024-05-04 19:43:18.311",
  "error": null
}'
+ echo '{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "aa4d62d1-e9c2-4bbb-b784-a2e666a98a39",
  "state": "normal",
  "checkpoint_tso": 449530560069238786,
  "checkpoint_time": "2024-05-04 19:43:18.311",
  "error": null
}'
{
  "upstream_id": 7365108570950873885,
  "namespace": "default",
  "id": "aa4d62d1-e9c2-4bbb-b784-a2e666a98a39",
  "state": "normal",
  "checkpoint_tso": 449530560069238786,
  "checkpoint_time": "2024-05-04 19:43:18.311",
  "error": null
}
++ echo '{' '"upstream_id":' 7365108570950873885, '"namespace":' '"default",' '"id":' '"aa4d62d1-e9c2-4bbb-b784-a2e666a98a39",' '"state":' '"normal",' '"checkpoint_tso":' 449530560069238786, '"checkpoint_time":' '"2024-05-04' '19:43:18.311",' '"error":' null '}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \n\o\r\m\a\l ]]
++ echo '{' '"upstream_id":' 7365108570950873885, '"namespace":' '"default",' '"id":' '"aa4d62d1-e9c2-4bbb-b784-a2e666a98a39",' '"state":' '"normal",' '"checkpoint_tso":' 449530560069238786, '"checkpoint_time":' '"2024-05-04' '19:43:18.311",' '"error":' null '}'
++ jq -r .error.message
+ message=null
+ [[ ! null =~ null ]]
run task successfully
check_safepoint_equal http://127.0.0.1:2379 7365108570950873885
check diff successfully
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
wait process cdc.test exit for 1-th time...
table ddl_manager.finish_mark not exists for 55-th check, retry later
wait process cdc.test exit for 2-th time...
cdc.test: no process found
wait process cdc.test exit for 3-th time...
process cdc.test already exit
[Sat May  4 19:43:21 CST 2024] <<<<<< run test case kafka_sink_error_resume success! >>>>>>
run task successfully
Changefeed remove successfully.
ID: e08dd9ed-4bf7-4060-8e52-d1d52c1d6226
CheckpointTs: 449530558378672132
SinkURI: kafka://127.0.0.1:9092/ticdc-gc-safepoint-15014?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760
check_safepoint_forward http://127.0.0.1:2379 7365108570950873885 449530560737968132 449530560069238786 449530558378672132
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table ddl_manager.finish_mark not exists for 56-th check, retry later
run task successfully
Changefeed remove successfully.
ID: aa4d62d1-e9c2-4bbb-b784-a2e666a98a39
CheckpointTs: 449530561262256133
SinkURI: kafka://127.0.0.1:9092/ticdc-gc-safepoint-15014?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760
check_safepoint_cleared http://127.0.0.1:2379 7365108570950873885
run task successfully
wait process cdc.test exit for 1-th time...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da9f60c000f	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44, pid:2162, start at 2024-05-04 19:43:23.541023258 +0800 CST m=+5.310376375	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:23.549 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:23.523 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:23.523 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
wait process cdc.test exit for 2-th time...
table ddl_manager.finish_mark not exists for 57-th check, retry later
wait process cdc.test exit for 3-th time...
cdc.test: no process found
wait process cdc.test exit for 4-th time...
process cdc.test already exit
[Sat May  4 19:43:25 CST 2024] <<<<<< run test case gc_safepoint success! >>>>>>
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da9f60c000f	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44, pid:2162, start at 2024-05-04 19:43:23.541023258 +0800 CST m=+5.310376375	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:23.549 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:23.523 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:23.523 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0da9f6200014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44, pid:2250, start at 2024-05-04 19:43:23.551684448 +0800 CST m=+5.261725809	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:23.559 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:23.528 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:23.528 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/kafka_big_messages_v2/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/kafka_big_messages_v2/tiflash/log/error.log
arg matches is ArgMatches { args: {"data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/kafka_big_messages_v2/tiflash/db/proxy"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/kafka_big_messages_v2/tiflash-proxy.toml"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/kafka_big_messages_v2/tiflash/log/proxy.log"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
table ddl_manager.finish_mark not exists for 58-th check, retry later
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/changefeed_reconstruct/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 7.53 secs (495182521 bytes/sec)
[Pipeline] {
[Pipeline] cache
table ddl_manager.finish_mark not exists for 59-th check, retry later
[Sat May  4 19:43:28 CST 2024] <<<<<< START cdc server in kafka_big_messages_v2 case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ GO_FAILPOINTS=
+ (( i = 0 ))
+ (( i <= 50 ))
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.kafka_big_messages_v2.36183620.out server --log-file /tmp/tidb_cdc_test/kafka_big_messages_v2/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/kafka_big_messages_v2/cdc_data --cluster-id default
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
start tidb cluster in /tmp/tidb_cdc_test/changefeed_reconstruct
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
table ddl_manager.finish_mark not exists for 60-th check, retry later
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:43:31 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/31d7b9ca-9c82-4ff9-857d-615b11d50d4b
	{"id":"31d7b9ca-9c82-4ff9-857d-615b11d50d4b","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823009}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436a52e9cb
	31d7b9ca-9c82-4ff9-857d-615b11d50d4b

/tidb/cdc/default/default/upstream/7365108684718240222
	{"id":7365108684718240222,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/31d7b9ca-9c82-4ff9-857d-615b11d50d4b
	{"id":"31d7b9ca-9c82-4ff9-857d-615b11d50d4b","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823009}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436a52e9cb
	31d7b9ca-9c82-4ff9-857d-615b11d50d4b

/tidb/cdc/default/default/upstream/7365108684718240222
	{"id":7365108684718240222,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/31d7b9ca-9c82-4ff9-857d-615b11d50d4b
	{"id":"31d7b9ca-9c82-4ff9-857d-615b11d50d4b","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823009}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436a52e9cb
	31d7b9ca-9c82-4ff9-857d-615b11d50d4b

/tidb/cdc/default/default/upstream/7365108684718240222
	{"id":7365108684718240222,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
Create changefeed successfully!
ID: 93a3eeeb-48db-4ea0-b9ee-0c4eaf2d08fd
Info: {"upstream_id":7365108684718240222,"namespace":"default","id":"93a3eeeb-48db-4ea0-b9ee-0c4eaf2d08fd","sink_uri":"kafka://127.0.0.1:9092/big-message-test?protocol=open-protocol\u0026partition-num=1\u0026kafka-version=2.4.1\u0026max-message-bytes=12582912","create_time":"2024-05-04T19:43:32.012014341+08:00","start_ts":449530562785837058,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":true,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530562785837058,"checkpoint_ts":449530562785837058,"checkpoint_time":"2024-05-04 19:43:28.674"}
[Sat May  4 19:43:32 CST 2024] <<<<<< START kafka consumer in kafka_big_messages_v2 case >>>>>>
Starting generate kafka big messages...
go: downloading github.com/pingcap/errors v0.11.5-0.20240318064555-6bd07397691f
go: downloading go.uber.org/atomic v1.11.0
table ddl_manager.finish_mark not exists for 61-th check, retry later
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/mq_sink_lost_callback/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:43:32 CST 2024] <<<<<< run test case mq_sink_lost_callback success! >>>>>>
table ddl_manager.finish_mark not exists for 62-th check, retry later
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/mq_sink_dispatcher/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
table ddl_manager.finish_mark not exists for 63-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table kafka_big_messages.test not exists for 1-th check, retry later
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/changefeed_pause_resume/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
table ddl_manager.finish_mark not exists for 64-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table kafka_big_messages.test exists
start tidb cluster in /tmp/tidb_cdc_test/mq_sink_dispatcher
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
check diff failed 1-th time, retry later
table ddl_manager.finish_mark not exists for 65-th check, retry later
check diff failed 1-th time, retry later
check diff successfully
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
wait process cdc.test exit for 1-th time...
check diff successfully
wait process cdc.test exit for 2-th time...
wait process cdc.test exit for 1-th time...
cdc.test: no process found
wait process cdc.test exit for 3-th time...
process cdc.test already exit
[Sat May  4 19:43:42 CST 2024] <<<<<< run test case kv_client_stream_reconnect success! >>>>>>
table sink_retry.finish_mark_1 exists
check diff successfully
***************** properties *****************
"readproportion"="0"
"operationcount"="0"
"updateproportion"="0"
"readallfields"="true"
"threadcount"="2"
"dotransactions"="false"
"workload"="core"
"scanproportion"="0"
"mysql.port"="4000"
"recordcount"="10"
"mysql.user"="root"
"mysql.host"="127.0.0.1"
"mysql.db"="sink_retry"
"insertproportion"="0"
"requestdistribution"="uniform"
**********************************************
Run finished, takes 11.591607ms
INSERT - Takes(s): 0.0, Count: 10, OPS: 2753.3, Avg(us): 2227, Min(us): 509, Max(us): 7940, 95th(us): 8000, 99th(us): 8000
wait process cdc.test exit for 2-th time...
table ddl_manager.finish_mark not exists for 66-th check, retry later
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0daaf9880013	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf, pid:4468, start at 2024-05-04 19:43:40.161596722 +0800 CST m=+5.340607014	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:40.169 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:40.130 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:40.130 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0daaf9880013	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf, pid:4468, start at 2024-05-04 19:43:40.161596722 +0800 CST m=+5.340607014	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:40.169 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:40.130 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:40.130 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0daaf9640016	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf, pid:4557, start at 2024-05-04 19:43:40.165932191 +0800 CST m=+5.296172257	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:40.173 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:40.171 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:40.171 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/changefeed_reconstruct/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/changefeed_reconstruct/tiflash/log/error.log
arg matches is ArgMatches { args: {"engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/changefeed_reconstruct/tiflash/log/proxy.log"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/changefeed_reconstruct/tiflash-proxy.toml"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/changefeed_reconstruct/tiflash/db/proxy"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
wait process cdc.test exit for 3-th time...
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
cdc.test: no process found
wait process cdc.test exit for 4-th time...
process cdc.test already exit
[Sat May  4 19:43:43 CST 2024] <<<<<< run test case kafka_big_messages_v2 success! >>>>>>
table ddl_manager.finish_mark not exists for 67-th check, retry later
Starting Upstream TiDB...
[Sat May  4 19:43:45 CST 2024] <<<<<< START cdc server in changefeed_reconstruct case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ GO_FAILPOINTS=
+ (( i = 0 ))
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.changefeed_reconstruct.59675969.out server --log-file /tmp/tidb_cdc_test/changefeed_reconstruct/cdcserver1.log --log-level debug --data-dir /tmp/tidb_cdc_test/changefeed_reconstruct/cdc_dataserver1 --cluster-id default --addr 127.0.0.1:8300 --pd http://127.0.0.1:2379
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
start tidb cluster in /tmp/tidb_cdc_test/changefeed_pause_resume
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
table ddl_manager.finish_mark not exists for 68-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:43:48 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/689e041c-68a7-429a-adfb-c35063a6024d
	{"id":"689e041c-68a7-429a-adfb-c35063a6024d","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823025}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436a95bdcc
	689e041c-68a7-429a-adfb-c35063a6024d

/tidb/cdc/default/default/upstream/7365108752870128665
	{"id":7365108752870128665,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/689e041c-68a7-429a-adfb-c35063a6024d
	{"id":"689e041c-68a7-429a-adfb-c35063a6024d","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823025}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436a95bdcc
	689e041c-68a7-429a-adfb-c35063a6024d

/tidb/cdc/default/default/upstream/7365108752870128665
	{"id":7365108752870128665,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/689e041c-68a7-429a-adfb-c35063a6024d
	{"id":"689e041c-68a7-429a-adfb-c35063a6024d","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823025}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436a95bdcc
	689e041c-68a7-429a-adfb-c35063a6024d

/tidb/cdc/default/default/upstream/7365108752870128665
	{"id":7365108752870128665,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
[Sat May  4 19:43:48 CST 2024] <<<<<< START kafka consumer in changefeed_reconstruct case >>>>>>
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 17.84 secs (208975644 bytes/sec)
[Pipeline] {
[Pipeline] cache
table ddl_manager.finish_mark not exists for 69-th check, retry later
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
***************** properties *****************
"mysql.host"="127.0.0.1"
"operationcount"="0"
"readallfields"="true"
"scanproportion"="0"
"readproportion"="0"
"workload"="core"
"requestdistribution"="uniform"
"mysql.db"="changefeed_reconstruct"
"threadcount"="4"
"mysql.port"="4000"
"mysql.user"="root"
"insertproportion"="0"
"recordcount"="50"
"updateproportion"="0"
"dotransactions"="false"
**********************************************
Run finished, takes 18.308788ms
INSERT - Takes(s): 0.0, Count: 48, OPS: 3578.5, Avg(us): 1453, Min(us): 953, Max(us): 4833, 95th(us): 5000, 99th(us): 5000
table changefeed_reconstruct.usertable not exists for 1-th check, retry later
table ddl_manager.finish_mark not exists for 70-th check, retry later
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dab98380019	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:12371, start at 2024-05-04 19:43:50.321286139 +0800 CST m=+5.147268184	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:50.329 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:50.286 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:50.286 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dab98380019	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:12371, start at 2024-05-04 19:43:50.321286139 +0800 CST m=+5.147268184	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:50.329 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:50.286 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:50.286 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dab9820000e	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:12446, start at 2024-05-04 19:43:50.295704033 +0800 CST m=+5.066163109	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:50.303 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:50.280 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:50.280 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/mq_sink_dispatcher/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/mq_sink_dispatcher/tiflash/log/error.log
arg matches is ArgMatches { args: {"engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/mq_sink_dispatcher/tiflash-proxy.toml"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/mq_sink_dispatcher/tiflash/db/proxy"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/mq_sink_dispatcher/tiflash/log/proxy.log"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
table changefeed_reconstruct.usertable exists
check diff failed 1-th time, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table ddl_manager.finish_mark not exists for 71-th check, retry later
[Sat May  4 19:43:53 CST 2024] <<<<<< START cdc server in mq_sink_dispatcher case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ GO_FAILPOINTS=
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.mq_sink_dispatcher.1389013892.out server --log-file /tmp/tidb_cdc_test/mq_sink_dispatcher/cdc.log --log-level info --data-dir /tmp/tidb_cdc_test/mq_sink_dispatcher/cdc_data --cluster-id default
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
check diff successfully
wait process 5972 exit for 1-th time...
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/cdc/run.sh using Sink-Type: kafka... <<=================
wait process 5972 exit for 2-th time...
table ddl_manager.finish_mark exists
check diff successfully
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils/kill_cdc_pid: line 19: kill: (5972) - No such process
wait process 5972 exit for 3-th time...
process 5972 already exit
check_no_capture http://127.0.0.1:2379
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
parse error: Invalid numeric literal at line 1, column 6
run task successfully
[Sat May  4 19:43:55 CST 2024] <<<<<< START cdc server in changefeed_reconstruct case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ GO_FAILPOINTS=
+ (( i = 0 ))
+ (( i <= 50 ))
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.changefeed_reconstruct.62876289.out server --log-file /tmp/tidb_cdc_test/changefeed_reconstruct/cdcserver2.log --log-level debug --data-dir /tmp/tidb_cdc_test/changefeed_reconstruct/cdc_dataserver2 --cluster-id default --addr 127.0.0.1:8300
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
wait process cdc.test exit for 1-th time...
wait process cdc.test exit for 2-th time...
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/multi_tables_ddl_v2/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
wait process cdc.test exit for 3-th time...
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:43:56 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/13eb1918-a954-4659-9460-30738467b079
	{"id":"13eb1918-a954-4659-9460-30738467b079","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823033}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436aba4bc1
	13eb1918-a954-4659-9460-30738467b079

/tidb/cdc/default/default/upstream/7365108798475769246
	{"id":7365108798475769246,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/13eb1918-a954-4659-9460-30738467b079
	{"id":"13eb1918-a954-4659-9460-30738467b079","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823033}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436aba4bc1
	13eb1918-a954-4659-9460-30738467b079

/tidb/cdc/default/default/upstream/7365108798475769246
	{"id":7365108798475769246,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/13eb1918-a954-4659-9460-30738467b079
	{"id":"13eb1918-a954-4659-9460-30738467b079","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823033}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436aba4bc1
	13eb1918-a954-4659-9460-30738467b079

/tidb/cdc/default/default/upstream/7365108798475769246
	{"id":7365108798475769246,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.mq_sink_dispatcher.cli.13964.out cli tso query --pd=http://127.0.0.1:2379
cdc.test: no process found
wait process cdc.test exit for 4-th time...
process cdc.test already exit
[Sat May  4 19:43:56 CST 2024] <<<<<< run test case ddl_manager success! >>>>>>
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
The 1 times to try to start tidb cluster...
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 6.81 secs (547226458 bytes/sec)
[Pipeline] {
[Pipeline] cache
+ set +x
+ tso='449530570233872388
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530570233872388 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.mq_sink_dispatcher.cli.14007.out cli changefeed create --start-ts=449530570233872388 '--sink-uri=kafka://127.0.0.1:9092/dispatcher-test?protocol=canal-json&enable-tidb-extension=true' -c test --config=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/mq_sink_dispatcher/conf/changefeed.toml
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:43:58 GMT
< Content-Type: text/plain; charset=utf-8
< Transfer-Encoding: chunked
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:

changefeedID: default/0bce12cd-4eb2-449c-9031-ca8878fec6ed
{UpstreamID:7365108752870128665 Namespace:default ID:0bce12cd-4eb2-449c-9031-ca8878fec6ed SinkURI:kafka://127.0.0.1:9092/ticdc-changefeed-reconstruct-14900?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760 CreateTime:2024-05-04 19:43:48.436272066 +0800 CST StartTs:449530567925694471 TargetTs:0 AdminJobType:noop Engine:unified SortDir: Config:0xc0017583f0 State:normal Error:<nil> Warning:<nil> CreatorVersion:v8.2.0-alpha-88-g57a3c8e82 Epoch:449530567965016066}
{CheckpointTs:449530569131556868 MinTableBarrierTs:449530570442276869 AdminJobType:noop}
span: {table_id:106,start_key:7480000000000000ff6a5f720000000000fa,end_key:7480000000000000ff6a5f730000000000fa}, resolvedTs: 449530569131556868, checkpointTs: 449530569131556868, state: Preparing



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca
	{"id":"4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823035}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436a95be73
	4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca

/tidb/cdc/default/default/changefeed/info/0bce12cd-4eb2-449c-9031-ca8878fec6ed
	{"upstream-id":7365108752870128665,"namespace":"default","changefeed-id":"0bce12cd-4eb2-449c-9031-ca8878fec6ed","sink-uri":"kafka://127.0.0.1:9092/ticdc-changefeed-reconstruct-14900?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create-time":"2024-05-04T19:43:48.436272066+08:00","start-ts":449530567925694471,"target-ts":0,"admin-job-type":0,"sort-engine":"","sort-dir":"","config":{"memory-quota":1073741824,"case-sensitive":false,"force-replicate":false,"check-gc-safe-point":true,"enable-sync-point":false,"enable-table-monitor":false,"ignore-ineligible-table":false,"bdr-mode":false,"sync-point-interval":600000000000,"sync-point-retention":86400000000000,"filter":{"rules":["*.*"],"ignore-txn-start-ts":null,"event-filters":null},"mounter":{"worker-num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include-commit-ts":false,"binary-encoding-method":"base64","output-old-value":false,"output-handle-key":false},"encoder-concurrency":32,"terminator":"\r\n","date-separator":"day","enable-partition-separator":true,"enable-kafka-sink-v2":false,"only-output-updated-columns":false,"delete-only-output-handle-key-columns":false,"content-compatible":false,"advance-timeout-in-sec":150,"send-bootstrap-interval-in-sec":120,"send-bootstrap-in-msg-count":10000,"send-bootstrap-to-all-partition":true,"debezium-disable-schema":false,"open":{"output-old-value":true},"debezium":{"output-old-value":true}},"consistent":{"level":"none","max-log-size":64,"flush-interval":2000,"meta-flush-interval":200,"encoding-worker-num":16,"flush-worker-num":8,"storage":"","use-file-backend":false,"compression":"","memory-usage":{"memory-quota-percentage":50}},"scheduler":{"enable-table-across-nodes":false,"region-threshold":100000,"write-key-threshold":0,"region-per-span":0},"integrity":{"integrity-check-level":"none","corruption-handle-level":"warn"},"changefeed-error-stuck-duration":1800000000000,"synced-status":{"synced-check-interval":300,"checkpoint-interval":15},"sql-mode":""},"state":"normal","error":null,"warning":null,"creator-version":"v8.2.0-alpha-88-g57a3c8e82","epoch":449530567965016066}

/tidb/cdc/default/default/changefeed/status/0bce12cd-4eb2-449c-9031-ca8878fec6ed
	{"checkpoint-ts":449530569131556868,"min-table-barrier-ts":449530570442276869,"admin-job-type":0}

/tidb/cdc/default/default/task/position/4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca/0bce12cd-4eb2-449c-9031-ca8878fec6ed
	{"checkpoint-ts":0,"resolved-ts":0,"count":0,"error":null,"warning":null}

/tidb/cdc/default/default/upstream/7365108752870128665
	{"id":7365108752870128665,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:

changefeedID: default/0bce12cd-4eb2-449c-9031-ca8878fec6ed
{UpstreamID:7365108752870128665 Namespace:default ID:0bce12cd-4eb2-449c-9031-ca8878fec6ed SinkURI:kafka://127.0.0.1:9092/ticdc-changefeed-reconstruct-14900?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760 CreateTime:2024-05-04 19:43:48.436272066 +0800 CST StartTs:449530567925694471 TargetTs:0 AdminJobType:noop Engine:unified SortDir: Config:0xc0017583f0 State:normal Error:<nil> Warning:<nil> CreatorVersion:v8.2.0-alpha-88-g57a3c8e82 Epoch:449530567965016066}
{CheckpointTs:449530569131556868 MinTableBarrierTs:449530570442276869 AdminJobType:noop}
span: {table_id:106,start_key:7480000000000000ff6a5f720000000000fa,end_key:7480000000000000ff6a5f730000000000fa}, resolvedTs: 449530569131556868, checkpointTs: 449530569131556868, state: Preparing



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca
	{"id":"4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823035}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436a95be73
	4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca

/tidb/cdc/default/default/changefeed/info/0bce12cd-4eb2-449c-9031-ca8878fec6ed
	{"upstream-id":7365108752870128665,"namespace":"default","changefeed-id":"0bce12cd-4eb2-449c-9031-ca8878fec6ed","sink-uri":"kafka://127.0.0.1:9092/ticdc-changefeed-reconstruct-14900?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create-time":"2024-05-04T19:43:48.436272066+08:00","start-ts":449530567925694471,"target-ts":0,"admin-job-type":0,"sort-engine":"","sort-dir":"","config":{"memory-quota":1073741824,"case-sensitive":false,"force-replicate":false,"check-gc-safe-point":true,"enable-sync-point":false,"enable-table-monitor":false,"ignore-ineligible-table":false,"bdr-mode":false,"sync-point-interval":600000000000,"sync-point-retention":86400000000000,"filter":{"rules":["*.*"],"ignore-txn-start-ts":null,"event-filters":null},"mounter":{"worker-num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include-commit-ts":false,"binary-encoding-method":"base64","output-old-value":false,"output-handle-key":false},"encoder-concurrency":32,"terminator":"\r\n","date-separator":"day","enable-partition-separator":true,"enable-kafka-sink-v2":false,"only-output-updated-columns":false,"delete-only-output-handle-key-columns":false,"content-compatible":false,"advance-timeout-in-sec":150,"send-bootstrap-interval-in-sec":120,"send-bootstrap-in-msg-count":10000,"send-bootstrap-to-all-partition":true,"debezium-disable-schema":false,"open":{"output-old-value":true},"debezium":{"output-old-value":true}},"consistent":{"level":"none","max-log-size":64,"flush-interval":2000,"meta-flush-interval":200,"encoding-worker-num":16,"flush-worker-num":8,"storage":"","use-file-backend":false,"compression":"","memory-usage":{"memory-quota-percentage":50}},"scheduler":{"enable-table-across-nodes":false,"region-threshold":100000,"write-key-threshold":0,"region-per-span":0},"integrity":{"integrity-check-level":"none","corruption-handle-level":"warn"},"changefeed-error-stuck-duration":1800000000000,"synced-status":{"synced-check-interval":300,"checkpoint-interval":15},"sql-mode":""},"state":"normal","error":null,"warning":null,"creator-version":"v8.2.0-alpha-88-g57a3c8e82","epoch":449530567965016066}

/tidb/cdc/default/default/changefeed/status/0bce12cd-4eb2-449c-9031-ca8878fec6ed
	{"checkpoint-ts":449530569131556868,"min-table-barrier-ts":449530570442276869,"admin-job-type":0}

/tidb/cdc/default/default/task/position/4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca/0bce12cd-4eb2-449c-9031-ca8878fec6ed
	{"checkpoint-ts":0,"resolved-ts":0,"count":0,"error":null,"warning":null}

/tidb/cdc/default/default/upstream/7365108752870128665
	{"id":7365108752870128665,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:

changefeedID: default/0bce12cd-4eb2-449c-9031-ca8878fec6ed
{UpstreamID:7365108752870128665 Namespace:default ID:0bce12cd-4eb2-449c-9031-ca8878fec6ed SinkURI:kafka://127.0.0.1:9092/ticdc-changefeed-reconstruct-14900?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760 CreateTime:2024-05-04 19:43:48.436272066 +0800 CST StartTs:449530567925694471 TargetTs:0 AdminJobType:noop Engine:unified SortDir: Config:0xc0017583f0 State:normal Error:<nil> Warning:<nil> CreatorVersion:v8.2.0-alpha-88-g57a3c8e82 Epoch:449530567965016066}
{CheckpointTs:449530569131556868 MinTableBarrierTs:449530570442276869 AdminJobType:noop}
span: {table_id:106,start_key:7480000000000000ff6a5f720000000000fa,end_key:7480000000000000ff6a5f730000000000fa}, resolvedTs: 449530569131556868, checkpointTs: 449530569131556868, state: Preparing



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca
	{"id":"4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823035}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436a95be73
	4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca

/tidb/cdc/default/default/changefeed/info/0bce12cd-4eb2-449c-9031-ca8878fec6ed
	{"upstream-id":7365108752870128665,"namespace":"default","changefeed-id":"0bce12cd-4eb2-449c-9031-ca8878fec6ed","sink-uri":"kafka://127.0.0.1:9092/ticdc-changefeed-reconstruct-14900?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create-time":"2024-05-04T19:43:48.436272066+08:00","start-ts":449530567925694471,"target-ts":0,"admin-job-type":0,"sort-engine":"","sort-dir":"","config":{"memory-quota":1073741824,"case-sensitive":false,"force-replicate":false,"check-gc-safe-point":true,"enable-sync-point":false,"enable-table-monitor":false,"ignore-ineligible-table":false,"bdr-mode":false,"sync-point-interval":600000000000,"sync-point-retention":86400000000000,"filter":{"rules":["*.*"],"ignore-txn-start-ts":null,"event-filters":null},"mounter":{"worker-num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include-commit-ts":false,"binary-encoding-method":"base64","output-old-value":false,"output-handle-key":false},"encoder-concurrency":32,"terminator":"\r\n","date-separator":"day","enable-partition-separator":true,"enable-kafka-sink-v2":false,"only-output-updated-columns":false,"delete-only-output-handle-key-columns":false,"content-compatible":false,"advance-timeout-in-sec":150,"send-bootstrap-interval-in-sec":120,"send-bootstrap-in-msg-count":10000,"send-bootstrap-to-all-partition":true,"debezium-disable-schema":false,"open":{"output-old-value":true},"debezium":{"output-old-value":true}},"consistent":{"level":"none","max-log-size":64,"flush-interval":2000,"meta-flush-interval":200,"encoding-worker-num":16,"flush-worker-num":8,"storage":"","use-file-backend":false,"compression":"","memory-usage":{"memory-quota-percentage":50}},"scheduler":{"enable-table-across-nodes":false,"region-threshold":100000,"write-key-threshold":0,"region-per-span":0},"integrity":{"integrity-check-level":"none","corruption-handle-level":"warn"},"changefeed-error-stuck-duration":1800000000000,"synced-status":{"synced-check-interval":300,"checkpoint-interval":15},"sql-mode":""},"state":"normal","error":null,"warning":null,"creator-version":"v8.2.0-alpha-88-g57a3c8e82","epoch":449530567965016066}

/tidb/cdc/default/default/changefeed/status/0bce12cd-4eb2-449c-9031-ca8878fec6ed
	{"checkpoint-ts":449530569131556868,"min-table-barrier-ts":449530570442276869,"admin-job-type":0}

/tidb/cdc/default/default/task/position/4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca/0bce12cd-4eb2-449c-9031-ca8878fec6ed
	{"checkpoint-ts":0,"resolved-ts":0,"count":0,"error":null,"warning"+ grep -q 'etcd info'
:null}

/tidb/cdc/default/default/upstream/7365108752870128665
	{"id":7365108752870128665,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ break
+ set +x
cdc.test cli capture list --pd=http://127.0.0.1:2379 2>&1 | grep id
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dabfb300012	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk, pid:4802, start at 2024-05-04 19:43:56.650710913 +0800 CST m=+5.778211786	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:56.656 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:56.620 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:56.620 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dabfb300012	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk, pid:4802, start at 2024-05-04 19:43:56.650710913 +0800 CST m=+5.778211786	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:56.656 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:56.620 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:56.620 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dabf3540009	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gzt73-t51wk, pid:4878, start at 2024-05-04 19:43:56.125755683 +0800 CST m=+5.203559753	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:45:56.132 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:43:56.117 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:33:56.117 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/changefeed_pause_resume/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/changefeed_pause_resume/tiflash/log/error.log
arg matches is ArgMatches { args: {"engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/changefeed_pause_resume/tiflash/log/proxy.log"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/changefeed_pause_resume/tiflash/db/proxy"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/changefeed_pause_resume/tiflash-proxy.toml"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
Create changefeed successfully!
ID: test
Info: {"upstream_id":7365108798475769246,"namespace":"default","id":"test","sink_uri":"kafka://127.0.0.1:9092/dispatcher-test?protocol=canal-json\u0026enable-tidb-extension=true","create_time":"2024-05-04T19:43:59.014857819+08:00","start_ts":449530570233872388,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"canal-json","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"dispatchers":[{"matcher":["verify.t"],"partition":"index-value"},{"matcher":["dispatcher.index"],"partition":"index-value","index":"idx_a"}],"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530570233872388,"checkpoint_ts":449530570233872388,"checkpoint_time":"2024-05-04 19:43:57.086"}
PASS
coverage: 2.5% of statements in github.com/pingcap/tiflow/...
start tidb cluster in /tmp/tidb_cdc_test/multi_tables_ddl_v2
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
    "id": "4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca",
    "cluster-id": "default"
run task successfully
+ set +x
check_changefeed_state http://127.0.0.1:2379 test normal null
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=test
+ expected_state=normal
+ error_msg=null
+ tls_dir=null
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c test -s
+ info='{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530570233872388,
  "checkpoint_time": "2024-05-04 19:43:57.086",
  "error": null
}'
+ echo '{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530570233872388,
  "checkpoint_time": "2024-05-04 19:43:57.086",
  "error": null
}'
{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530570233872388,
  "checkpoint_time": "2024-05-04 19:43:57.086",
  "error": null
}
++ echo '{' '"upstream_id":' 7365108798475769246, '"namespace":' '"default",' '"id":' '"test",' '"state":' '"normal",' '"checkpoint_tso":' 449530570233872388, '"checkpoint_time":' '"2024-05-04' '19:43:57.086",' '"error":' null '}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \n\o\r\m\a\l ]]
++ echo '{' '"upstream_id":' 7365108798475769246, '"namespace":' '"default",' '"id":' '"test",' '"state":' '"normal",' '"checkpoint_tso":' 449530570233872388, '"checkpoint_time":' '"2024-05-04' '19:43:57.086",' '"error":' null '}'
++ jq -r .error.message
+ message=null
+ [[ ! null =~ null ]]
run task successfully
check_changefeed_state http://127.0.0.1:2379 test failed ErrDispatcherFailed
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=test
+ expected_state=failed
+ error_msg=ErrDispatcherFailed
+ tls_dir=ErrDispatcherFailed
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c test -s
+ info='{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530570233872388,
  "checkpoint_time": "2024-05-04 19:43:57.086",
  "error": null
}'
+ echo '{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530570233872388,
  "checkpoint_time": "2024-05-04 19:43:57.086",
  "error": null
}'
{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530570233872388,
  "checkpoint_time": "2024-05-04 19:43:57.086",
  "error": null
}
++ echo '{' '"upstream_id":' 7365108798475769246, '"namespace":' '"default",' '"id":' '"test",' '"state":' '"normal",' '"checkpoint_tso":' 449530570233872388, '"checkpoint_time":' '"2024-05-04' '19:43:57.086",' '"error":' null '}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \f\a\i\l\e\d ]]
+ echo 'changefeed state normal does not equal to failed'
changefeed state normal does not equal to failed
+ exit 1
run task failed 1-th time, retry later
[Sat May  4 19:44:01 CST 2024] <<<<<< START cdc server in changefeed_pause_resume case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ GO_FAILPOINTS=
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.changefeed_pause_resume.63376339.out server --log-file /tmp/tidb_cdc_test/changefeed_pause_resume/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/changefeed_pause_resume/cdc_data --cluster-id default --addr 127.0.0.1:8300 --pd http://127.0.0.1:2379
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
start tidb cluster in /tmp/tidb_cdc_test/cdc
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
capture_id: 4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca
check_processor_table_count http://127.0.0.1:2379 0bce12cd-4eb2-449c-9031-ca8878fec6ed 4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca 1
run task successfully
check_processor_table_count http://127.0.0.1:2379 0bce12cd-4eb2-449c-9031-ca8878fec6ed 4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca 0
table count 1 does equal to expected count 0
run task failed 1-th time, retry later
check_changefeed_state http://127.0.0.1:2379 test failed ErrDispatcherFailed
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=test
+ expected_state=failed
+ error_msg=ErrDispatcherFailed
+ tls_dir=ErrDispatcherFailed
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c test -s
+ info='{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530571204067370,
  "checkpoint_time": "2024-05-04 19:44:00.787",
  "error": null
}'
+ echo '{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530571204067370,
  "checkpoint_time": "2024-05-04 19:44:00.787",
  "error": null
}'
{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530571204067370,
  "checkpoint_time": "2024-05-04 19:44:00.787",
  "error": null
}
++ echo '{' '"upstream_id":' 7365108798475769246, '"namespace":' '"default",' '"id":' '"test",' '"state":' '"normal",' '"checkpoint_tso":' 449530571204067370, '"checkpoint_time":' '"2024-05-04' '19:44:00.787",' '"error":' null '}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \f\a\i\l\e\d ]]
+ echo 'changefeed state normal does not equal to failed'
changefeed state normal does not equal to failed
+ exit 1
run task failed 2-th time, retry later
Verifying downstream PD is started...
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:44:04 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/f59e4dd0-c4cd-46f3-911b-e38a9b2335bc
	{"id":"f59e4dd0-c4cd-46f3-911b-e38a9b2335bc","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823041}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436ad466ca
	f59e4dd0-c4cd-46f3-911b-e38a9b2335bc

/tidb/cdc/default/default/upstream/7365108823396415104
	{"id":7365108823396415104,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/f59e4dd0-c4cd-46f3-911b-e38a9b2335bc
	{"id":"f59e4dd0-c4cd-46f3-911b-e38a9b2335bc","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823041}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436ad466ca
	f59e4dd0-c4cd-46f3-911b-e38a9b2335bc

/tidb/cdc/default/default/upstream/7365108823396415104
	{"id":7365108823396415104,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/f59e4dd0-c4cd-46f3-911b-e38a9b2335bc
	{"id":"f59e4dd0-c4cd-46f3-911b-e38a9b2335bc","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823041}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436ad466ca
	f59e4dd0-c4cd-46f3-911b-e38a9b2335bc

/tidb/cdc/default/default/upstream/7365108823396415104
	{"id":7365108823396415104,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
check_processor_table_count http://127.0.0.1:2379 0bce12cd-4eb2-449c-9031-ca8878fec6ed 4fe11fdb-a7ed-427f-8bd8-dc4e7bdb52ca 0
[Sat May  4 19:44:04 CST 2024] <<<<<< START kafka consumer in changefeed_pause_resume case >>>>>>
run task successfully
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
***************** properties *****************
"requestdistribution"="uniform"
"workload"="core"
"mysql.user"="root"
"readproportion"="0"
"mysql.db"="changefeed_reconstruct"
"dotransactions"="false"
"recordcount"="50"
"scanproportion"="0"
"insertproportion"="0"
"mysql.port"="4000"
"updateproportion"="0"
"readallfields"="true"
"mysql.host"="127.0.0.1"
"operationcount"="0"
"threadcount"="4"
**********************************************
Run finished, takes 19.81051ms
INSERT - Takes(s): 0.0, Count: 48, OPS: 3112.6, Avg(us): 1588, Min(us): 1090, Max(us): 4635, 95th(us): 5000, 99th(us): 5000
table changefeed_reconstruct.usertable not exists for 1-th check, retry later
table changefeed_pause_resume.t1 not exists for 1-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
check_changefeed_state http://127.0.0.1:2379 test failed ErrDispatcherFailed
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=test
+ expected_state=failed
+ error_msg=ErrDispatcherFailed
+ tls_dir=ErrDispatcherFailed
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c test -s
+ info='{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "failed",
  "checkpoint_tso": 449530571204067370,
  "checkpoint_time": "2024-05-04 19:44:00.787",
  "error": {
    "time": "2024-05-04T19:44:04.552478027+08:00",
    "addr": "127.0.0.1:8300",
    "code": "CDC:ErrDispatcherFailed",
    "message": "[CDC:ErrDispatcherFailed]index not found when dispatch event, table: index, index: idx_a"
  }
}'
+ echo '{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "failed",
  "checkpoint_tso": 449530571204067370,
  "checkpoint_time": "2024-05-04 19:44:00.787",
  "error": {
    "time": "2024-05-04T19:44:04.552478027+08:00",
    "addr": "127.0.0.1:8300",
    "code": "CDC:ErrDispatcherFailed",
    "message": "[CDC:ErrDispatcherFailed]index not found when dispatch event, table: index, index: idx_a"
  }
}'
{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "failed",
  "checkpoint_tso": 449530571204067370,
  "checkpoint_time": "2024-05-04 19:44:00.787",
  "error": {
    "time": "2024-05-04T19:44:04.552478027+08:00",
    "addr": "127.0.0.1:8300",
    "code": "CDC:ErrDispatcherFailed",
    "message": "[CDC:ErrDispatcherFailed]index not found when dispatch event, table: index, index: idx_a"
  }
}
++ jq -r .state
++ echo '{' '"upstream_id":' 7365108798475769246, '"namespace":' '"default",' '"id":' '"test",' '"state":' '"failed",' '"checkpoint_tso":' 449530571204067370, '"checkpoint_time":' '"2024-05-04' '19:44:00.787",' '"error":' '{' '"time":' '"2024-05-04T19:44:04.552478027+08:00",' '"addr":' '"127.0.0.1:8300",' '"code":' '"CDC:ErrDispatcherFailed",' '"message":' '"[CDC:ErrDispatcherFailed]index' not found when dispatch event, table: index, index: 'idx_a"' '}' '}'
+ state=failed
+ [[ ! failed == \f\a\i\l\e\d ]]
++ echo '{' '"upstream_id":' 7365108798475769246, '"namespace":' '"default",' '"id":' '"test",' '"state":' '"failed",' '"checkpoint_tso":' 449530571204067370, '"checkpoint_time":' '"2024-05-04' '19:44:00.787",' '"error":' '{' '"time":' '"2024-05-04T19:44:04.552478027+08:00",' '"addr":' '"127.0.0.1:8300",' '"code":' '"CDC:ErrDispatcherFailed",' '"message":' '"[CDC:ErrDispatcherFailed]index' not found when dispatch event, table: index, index: 'idx_a"' '}' '}'
++ jq -r .error.message
+ message='[CDC:ErrDispatcherFailed]index not found when dispatch event, table: index, index: idx_a'
+ [[ ! [CDC:ErrDispatcherFailed]index not found when dispatch event, table: index, index: idx_a =~ ErrDispatcherFailed ]]
run task successfully
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.mq_sink_dispatcher.cli.14271.out cli changefeed update -c test '--sink-uri=kafka://127.0.0.1:9092/dispatcher-test?protocol=canal-json&enable-tidb-extension=true' --config=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/mq_sink_dispatcher/conf/new_changefeed.toml --no-confirm
Diff of changefeed config:
{Type:update Path:[Config SyncPointInterval] From:<nil> To:0xc003902038}
{Type:update Path:[Config SyncPointRetention] From:<nil> To:0xc003902048}
{Type:update Path:[Config Sink DispatchRules 0 Matcher 0] From:verify.t To:dispatcher.index}
{Type:delete Path:[Config Sink DispatchRules 1 Matcher 0] From:dispatcher.index To:<nil>}
{Type:delete Path:[Config Sink DispatchRules 1 PartitionRule] From:index-value To:<nil>}
{Type:delete Path:[Config Sink DispatchRules 1 IndexName] From:idx_a To:<nil>}
{Type:update Path:[Config Consistent] From:<nil> To:0xc001353880}
Update changefeed config successfully! 
ID: test
Info: {"upstream_id":7365108798475769246,"namespace":"default","id":"test","sink_uri":"kafka://127.0.0.1:9092/dispatcher-test?protocol=canal-json\u0026enable-tidb-extension=true","create_time":"2024-05-04T19:43:59.014857819+08:00","start_ts":449530570233872388,"admin_job_type":1,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_table_monitor":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"canal-json","dispatchers":[{"matcher":["dispatcher.index"],"partition":"index-value"}],"encoder_concurrency":32,"terminator":"\r\n","enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"failed","error":{"addr":"127.0.0.1:8300","code":"CDC:ErrDispatcherFailed","message":"[CDC:ErrDispatcherFailed]index not found when dispatch event, table: index, index: idx_a"},"creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":0,"checkpoint_ts":449530571204067370,"checkpoint_time":"2024-05-04 19:44:00.787"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
table changefeed_reconstruct.usertable exists
check diff failed 1-th time, retry later
table changefeed_pause_resume.t1 exists
table changefeed_pause_resume.t2 exists
table changefeed_pause_resume.t3 not exists for 1-th check, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.mq_sink_dispatcher.cli.14307.out cli changefeed resume -c test
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
PASS
check diff successfully
coverage: 2.1% of statements in github.com/pingcap/tiflow/...
wait process cdc.test exit for 1-th time...
table changefeed_pause_resume.t3 exists
wait process cdc.test exit for 2-th time...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
cdc.test: no process found
wait process cdc.test exit for 3-th time...
process cdc.test already exit
[Sat May  4 19:44:11 CST 2024] <<<<<< run test case changefeed_reconstruct success! >>>>>>
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dacd2900018	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44, pid:5926, start at 2024-05-04 19:44:10.447869879 +0800 CST m=+5.200901838	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:10.456 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:10.454 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:10.454 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dacd2900018	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44, pid:5926, start at 2024-05-04 19:44:10.447869879 +0800 CST m=+5.200901838	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:10.456 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:10.454 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:10.454 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dacd40c0015	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-xx2km-1zw44, pid:6008, start at 2024-05-04 19:44:10.527221061 +0800 CST m=+5.221794547	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:10.533 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:10.499 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:10.499 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/multi_tables_ddl_v2/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/multi_tables_ddl_v2/tiflash/log/error.log
arg matches is ArgMatches { args: {"data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/multi_tables_ddl_v2/tiflash/db/proxy"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/multi_tables_ddl_v2/tiflash-proxy.toml"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/multi_tables_ddl_v2/tiflash/log/proxy.log"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/multi_source/run.sh using Sink-Type: kafka... <<=================
+ set +x
check_changefeed_state http://127.0.0.1:2379 test normal null
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=test
+ expected_state=normal
+ error_msg=null
+ tls_dir=null
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c test -s
+ info='{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530571204067370,
  "checkpoint_time": "2024-05-04 19:44:00.787",
  "error": null
}'
+ echo '{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530571204067370,
  "checkpoint_time": "2024-05-04 19:44:00.787",
  "error": null
}'
{
  "upstream_id": 7365108798475769246,
  "namespace": "default",
  "id": "test",
  "state": "normal",
  "checkpoint_tso": 449530571204067370,
  "checkpoint_time": "2024-05-04 19:44:00.787",
  "error": null
}
++ echo '{' '"upstream_id":' 7365108798475769246, '"namespace":' '"default",' '"id":' '"test",' '"state":' '"normal",' '"checkpoint_tso":' 449530571204067370, '"checkpoint_time":' '"2024-05-04' '19:44:00.787",' '"error":' null '}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \n\o\r\m\a\l ]]
++ echo '{' '"upstream_id":' 7365108798475769246, '"namespace":' '"default",' '"id":' '"test",' '"state":' '"normal",' '"checkpoint_tso":' 449530571204067370, '"checkpoint_time":' '"2024-05-04' '19:44:00.787",' '"error":' null '}'
++ jq -r .error.message
+ message=null
+ [[ ! null =~ null ]]
run task successfully
table test.finish_mark not exists for 1-th check, retry later
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dacf2080008	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb, pid:4481, start at 2024-05-04 19:44:12.427204028 +0800 CST m=+5.236145328	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:12.434 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:12.418 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:12.418 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
[Sat May  4 19:44:13 CST 2024] <<<<<< START cdc server in multi_tables_ddl_v2 case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ GO_FAILPOINTS=
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.multi_tables_ddl_v2.74747476.out server --log-file /tmp/tidb_cdc_test/multi_tables_ddl_v2/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/multi_tables_ddl_v2/cdc_data --cluster-id default
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
The 1 times to try to start tidb cluster...
table sink_retry.finish_mark_2 exists
check diff successfully
wait process cdc.test exit for 1-th time...
wait process cdc.test exit for 2-th time...
wait process cdc.test exit for 3-th time...
cdc.test: no process found
wait process cdc.test exit for 4-th time...
process cdc.test already exit
[Sat May  4 19:44:13 CST 2024] <<<<<< run test case sink_retry success! >>>>>>
table test.finish_mark not exists for 2-th check, retry later
check diff failed 1-th time, retry later
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dacf2080008	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb, pid:4481, start at 2024-05-04 19:44:12.427204028 +0800 CST m=+5.236145328	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:12.434 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:12.418 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:12.418 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dacf3840007	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6tpkp-wg7xb, pid:4552, start at 2024-05-04 19:44:12.520494832 +0800 CST m=+5.274444690	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:12.527 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:12.513 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:12.513 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/cdc/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/cdc/tiflash/log/error.log
arg matches is ArgMatches { args: {"engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/cdc/tiflash/db/proxy"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/cdc/tiflash/log/proxy.log"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/cdc/tiflash-proxy.toml"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
table test.finish_mark exists
check diff successfully
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:44:16 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/99ead349-e665-4642-9374-422ccbf363e0
	{"id":"99ead349-e665-4642-9374-422ccbf363e0","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823054}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436b08aec4
	99ead349-e665-4642-9374-422ccbf363e0

/tidb/cdc/default/default/upstream/7365108887230533730
	{"id":7365108887230533730,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/99ead349-e665-4642-9374-422ccbf363e0
	{"id":"99ead349-e665-4642-9374-422ccbf363e0","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823054}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436b08aec4
	99ead349-e665-4642-9374-422ccbf363e0

/tidb/cdc/default/default/upstream/7365108887230533730
	{"id":7365108887230533730,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/99ead349-e665-4642-9374-422ccbf363e0
	{"id":"99ead349-e665-4642-9374-422ccbf363e0","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823054}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436b08aec4
	99ead349-e665-4642-9374-422ccbf363e0

/tidb/cdc/default/default/upstream/7365108887230533730
	{"id":7365108887230533730,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
wait process cdc.test exit for 1-th time...
Create changefeed successfully!
ID: test-normal
Info: {"upstream_id":7365108887230533730,"namespace":"default","id":"test-normal","sink_uri":"kafka://127.0.0.1:9092/ticdc-multi-tables-ddl-test-normal-11950?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:44:16.960358945+08:00","start_ts":449530574563966977,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["multi_tables_ddl_test.t1","multi_tables_ddl_test.t2","multi_tables_ddl_test.t3","multi_tables_ddl_test.t4","multi_tables_ddl_test.t1_7","multi_tables_ddl_test.t2_7","multi_tables_ddl_test.finish_mark"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":true,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530574563966977,"checkpoint_ts":449530574563966977,"checkpoint_time":"2024-05-04 19:44:13.604"}
check diff failed 2-th time, retry later
Create changefeed successfully!
ID: test-error-1
Info: {"upstream_id":7365108887230533730,"namespace":"default","id":"test-error-1","sink_uri":"kafka://127.0.0.1:9092/ticdc-multi-tables-ddl-test-error-1-24176?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:44:17.16010644+08:00","start_ts":449530574563966977,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["multi_tables_ddl_test.t5","multi_tables_ddl_test.t6","multi_tables_ddl_test.t7","multi_tables_ddl_test.t8"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":true,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530574563966977,"checkpoint_ts":449530574563966977,"checkpoint_time":"2024-05-04 19:44:13.604"}
wait process cdc.test exit for 2-th time...
Create changefeed successfully!
ID: test-error-2
Info: {"upstream_id":7365108887230533730,"namespace":"default","id":"test-error-2","sink_uri":"kafka://127.0.0.1:9092/ticdc-multi-tables-ddl-test-error-2-18572?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:44:17.356180224+08:00","start_ts":449530574563966977,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["multi_tables_ddl_test.t9","multi_tables_ddl_test.t10"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":true,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530574563966977,"checkpoint_ts":449530574563966977,"checkpoint_time":"2024-05-04 19:44:13.604"}
[Sat May  4 19:44:17 CST 2024] <<<<<< START kafka consumer in multi_tables_ddl_v2 case >>>>>>
[Sat May  4 19:44:17 CST 2024] <<<<<< START kafka consumer in multi_tables_ddl_v2 case >>>>>>
[Sat May  4 19:44:17 CST 2024] <<<<<< START kafka consumer in multi_tables_ddl_v2 case >>>>>>
[Sat May  4 19:44:17 CST 2024] <<<<<< START cdc server in cdc case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ GO_FAILPOINTS=
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.cdc.59285930.out server --log-file /tmp/tidb_cdc_test/cdc/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/cdc/cdc_data --cluster-id default
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
wait process cdc.test exit for 3-th time...
cdc.test: no process found
wait process cdc.test exit for 4-th time...
process cdc.test already exit
[Sat May  4 19:44:18 CST 2024] <<<<<< run test case mq_sink_dispatcher success! >>>>>>
check diff failed 3-th time, retry later
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:44:20 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/e6eb5158-f477-44b2-932f-648f2ec7efb1
	{"id":"e6eb5158-f477-44b2-932f-648f2ec7efb1","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823057}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436b1026d2
	e6eb5158-f477-44b2-932f-648f2ec7efb1

/tidb/cdc/default/default/upstream/7365108884341208501
	{"id":7365108884341208501,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/e6eb5158-f477-44b2-932f-648f2ec7efb1
	{"id":"e6eb5158-f477-44b2-932f-648f2ec7efb1","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823057}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436b1026d2
	e6eb5158-f477-44b2-932f-648f2ec7efb1

/tidb/cdc/default/default/upstream/7365108884341208501
	{"id":7365108884341208501,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/e6eb5158-f477-44b2-932f-648f2ec7efb1
	{"id":"e6eb5158-f477-44b2-932f-648f2ec7efb1","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823057}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436b1026d2
	e6eb5158-f477-44b2-932f-648f2ec7efb1

/tidb/cdc/default/default/upstream/7365108884341208501
	{"id":7365108884341208501,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.cdc.cli.5995.out cli changefeed create '--sink-uri=kafka://127.0.0.1:9092/ticdc-cdc-test-5808?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760' --config /tmp/tidb_cdc_test/cdc/pulsar_test.toml
Create changefeed successfully!
ID: ffe606f7-90d4-4cbc-8f24-b05f90668e02
Info: {"upstream_id":7365108884341208501,"namespace":"default","id":"ffe606f7-90d4-4cbc-8f24-b05f90668e02","sink_uri":"kafka://127.0.0.1:9092/ticdc-cdc-test-5808?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:44:21.095027443+08:00","start_ts":449530576481550342,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530576481550342,"checkpoint_ts":449530576481550342,"checkpoint_time":"2024-05-04 19:44:20.919"}
PASS
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
Cache restored successfully (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc)
3727185920 bytes in 21.03 secs (177217579 bytes/sec)
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] container
[Pipeline] {
check diff failed 4-th time, retry later
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 6 min 0 sec
[Pipeline] {
[Pipeline] sh
[Pipeline] sh
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
[Pipeline] sh
start tidb cluster in /tmp/tidb_cdc_test/multi_source
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
+ set +x
[Sat May  4 19:44:22 CST 2024] <<<<<< START kafka consumer in cdc case >>>>>>
[Pipeline] sh
go: downloading github.com/pingcap/log v1.1.1-0.20240314023424-862ccc32f18d
go: downloading github.com/pingcap/errors v0.11.5-0.20240318064555-6bd07397691f
go: downloading github.com/go-sql-driver/mysql v1.7.1
go: downloading go.uber.org/zap v1.27.0
go: downloading github.com/pingcap/tidb-tools v0.0.0-20240305021104-9f9bea84490b
go: downloading github.com/BurntSushi/toml v1.3.2
go: downloading github.com/pingcap/tidb v1.1.0-beta.0.20240415145106-cd9c676e9ba4
go: downloading github.com/pingcap/tidb/pkg/parser v0.0.0-20240410110152-5fc42c9be2f5
go: downloading go.uber.org/atomic v1.11.0
go: downloading gopkg.in/natefinch/lumberjack.v2 v2.2.1
go: downloading go.uber.org/multierr v1.11.0
go: downloading github.com/pingcap/failpoint v0.0.0-20220801062533-2eaa32854a6c
go: downloading google.golang.org/grpc v1.62.1
go: downloading github.com/coreos/go-semver v0.3.1
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
[Pipeline] sh
go: downloading github.com/golang/protobuf v1.5.4
go: downloading golang.org/x/net v0.24.0
go: downloading golang.org/x/sys v0.19.0
go: downloading google.golang.org/protobuf v1.33.0
go: downloading google.golang.org/genproto/googleapis/rpc v0.0.0-20240401170217-c3f982113cda
go: downloading google.golang.org/genproto v0.0.0-20240401170217-c3f982113cda
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
[Pipeline] sh
check diff failed 5-th time, retry later
go: downloading golang.org/x/text v0.14.0
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
Verifying downstream PD is started...
+ echo Waiting for zookeeper to be ready...
Waiting for zookeeper to be ready...
+ nc -z localhost 2181
+ echo Waiting for kafka to be ready...
Waiting for kafka to be ready...
+ nc -z localhost 9092
+ echo Waiting for kafka-broker to be ready...
Waiting for kafka-broker to be ready...
+ echo dump
+ nc localhost 2181
+ grep brokers
+ awk {$1=$1;print}
+ grep -F -w /brokers/ids/1
/brokers/ids/1
[Pipeline] // stage
[Pipeline] }
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/multi_capture/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
check diff successfully
go: downloading github.com/cznic/mathutil v0.0.0-20181122101859-297441e03548
go: downloading golang.org/x/exp v0.0.0-20240409090435-93d18d7e34b8
go: downloading golang.org/x/sync v0.7.0
go: downloading github.com/pingcap/tipb v0.0.0-20240318032315-55a7867ddd50
go: downloading github.com/tiancaiamao/gp v0.0.0-20221230034425-4025bc8a4d4a
go: downloading github.com/shirou/gopsutil/v3 v3.24.2
go: downloading go.etcd.io/etcd/client/v3 v3.5.12
go: downloading github.com/influxdata/tdigest v0.0.1
go: downloading github.com/pingcap/kvproto v0.0.0-20240227073058-929ab83f9754
go: downloading github.com/danjacques/gofslock v0.0.0-20240212154529-d899e02bfe22
go: downloading github.com/grpc-ecosystem/go-grpc-middleware v1.4.0
go: downloading github.com/cockroachdb/errors v1.11.1
go: downloading github.com/pingcap/sysutil v1.0.1-0.20240311050922-ae81ee01f3a5
go: downloading github.com/prometheus/client_golang v1.19.0
go: downloading github.com/opentracing/opentracing-go v1.2.0
go: downloading github.com/tikv/client-go/v2 v2.0.8-0.20240409022718-714958ccd4d5
go: downloading github.com/tikv/pd/client v0.0.0-20240322051414-fb9e2d561b6e
go: downloading github.com/jellydator/ttlcache/v3 v3.0.1
go: downloading github.com/opentracing/basictracer-go v1.1.0
go: downloading github.com/uber/jaeger-client-go v2.30.0+incompatible
go: downloading github.com/ngaut/pools v0.0.0-20180318154953-b7bc8c42aac7
go: downloading github.com/coocood/freecache v1.2.1
go: downloading gopkg.in/yaml.v2 v2.4.0
go: downloading github.com/prometheus/client_model v0.6.1
go: downloading github.com/twmb/murmur3 v1.1.6
go: downloading github.com/stretchr/testify v1.9.0
go: downloading github.com/docker/go-units v0.5.0
go: downloading github.com/spf13/pflag v1.0.5
go: downloading github.com/scalalang2/golang-fifo v0.1.5
go: downloading github.com/tidwall/btree v1.7.0
go: downloading cloud.google.com/go/storage v1.39.1
go: downloading github.com/google/uuid v1.6.0
go: downloading github.com/Azure/azure-sdk-for-go/sdk/azcore v1.9.1
go: downloading github.com/gogo/protobuf v1.3.2
go: downloading github.com/Azure/azure-sdk-for-go/sdk/azidentity v1.5.1
go: downloading github.com/Azure/azure-sdk-for-go/sdk/storage/azblob v1.0.0
go: downloading github.com/yangkeao/ldap/v3 v3.4.5-0.20230421065457-369a3bab1117
go: downloading golang.org/x/tools v0.20.0
go: downloading go.etcd.io/etcd/api/v3 v3.5.12
go: downloading github.com/aliyun/alibaba-cloud-sdk-go v1.61.1581
go: downloading github.com/tikv/pd v1.1.0-beta.0.20240407022249-7179657d129b
go: downloading github.com/aws/aws-sdk-go v1.50.0
go: downloading github.com/go-resty/resty/v2 v2.11.0
go: downloading github.com/klauspost/compress v1.17.8
go: downloading github.com/ks3sdklib/aws-sdk-go v1.2.9
go: downloading cloud.google.com/go v0.112.2
go: downloading golang.org/x/oauth2 v0.18.0
go: downloading google.golang.org/api v0.170.0
go: downloading github.com/dgraph-io/ristretto v0.1.1
go: downloading github.com/dolthub/swiss v0.2.1
go: downloading github.com/golang/snappy v0.0.4
go: downloading github.com/google/btree v1.1.2
go: downloading github.com/gorilla/mux v1.8.0
go: downloading github.com/ngaut/sync2 v0.0.0-20141008032647-7a24ed77b2ef
go: downloading github.com/cespare/xxhash/v2 v2.3.0
go: downloading go.uber.org/mock v0.4.0
go: downloading github.com/otiai10/copy v1.2.0
go: downloading github.com/asaskevich/govalidator v0.0.0-20230301143203-a9d515a09cc2
go: downloading github.com/cockroachdb/pebble v1.1.0
go: downloading github.com/jfcg/sorty/v2 v2.1.0
go: downloading github.com/lestrrat-go/jwx/v2 v2.0.21
go: downloading golang.org/x/time v0.5.0
go: downloading github.com/carlmjohnson/flagext v0.21.0
go: downloading github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec
go: downloading github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc
go: downloading gopkg.in/yaml.v3 v3.0.1
go: downloading github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2
go: downloading github.com/Azure/go-ntlmssp v0.0.0-20221128193559-754e69321358
go: downloading github.com/go-asn1-ber/asn1-ber v1.5.4
go: downloading go.etcd.io/etcd/client/pkg/v3 v3.5.12
go: downloading github.com/Azure/azure-sdk-for-go/sdk/internal v1.5.1
go: downloading github.com/AzureAD/microsoft-authentication-library-for-go v1.2.1
go: downloading golang.org/x/crypto v0.22.0
go: downloading github.com/beorn7/perks v1.0.1
go: downloading github.com/prometheus/common v0.52.2
go: downloading github.com/prometheus/procfs v0.13.0
go: downloading github.com/cockroachdb/logtags v0.0.0-20230118201751-21c54148d20b
go: downloading github.com/cockroachdb/redact v1.1.5
go: downloading github.com/getsentry/sentry-go v0.27.0
go: downloading github.com/pkg/errors v0.9.1
go: downloading github.com/uber/jaeger-lib v2.4.1+incompatible
go: downloading github.com/dgryski/go-farm v0.0.0-20200201041132-a6ae2369ad13
go: downloading github.com/dolthub/maphash v0.1.0
go: downloading github.com/cheggaaa/pb/v3 v3.0.8
go: downloading github.com/tklauser/go-sysconf v0.3.12
go: downloading github.com/joho/sqltocsv v0.0.0-20210428211105-a6d6801d59df
go: downloading github.com/jedib0t/go-pretty/v6 v6.2.2
go: downloading github.com/cloudfoundry/gosigar v1.3.6
go: downloading github.com/pingcap/badger v1.5.1-0.20230103063557-828f39b09b6d
go: downloading github.com/google/pprof v0.0.0-20240117000934-35fc243c5815
go: downloading github.com/spkg/bom v1.0.0
go: downloading github.com/xitongsys/parquet-go v1.6.0
go: downloading github.com/robfig/cron/v3 v3.0.1
go: downloading google.golang.org/genproto/googleapis/api v0.0.0-20240401170217-c3f982113cda
go: downloading github.com/jfcg/sixb v1.3.8
go: downloading github.com/robfig/cron v1.2.0
go: downloading github.com/wangjohn/quickselect v0.0.0-20161129230411-ed8402a42d5f
go: downloading cloud.google.com/go/compute/metadata v0.2.3
go: downloading cloud.google.com/go/iam v1.1.7
go: downloading github.com/googleapis/gax-go/v2 v2.12.3
go: downloading cloud.google.com/go/compute v1.25.1
go: downloading github.com/kr/pretty v0.3.1
go: downloading github.com/pingcap/goleveldb v0.0.0-20191226122134-f82aafb29989
go: downloading github.com/coreos/go-systemd/v22 v22.5.0
go: downloading github.com/VividCortex/ewma v1.2.0
go: downloading github.com/fatih/color v1.16.0
go: downloading github.com/mattn/go-colorable v0.1.13
go: downloading github.com/mattn/go-isatty v0.0.20
go: downloading github.com/mattn/go-runewidth v0.0.15
go: downloading github.com/kylelemons/godebug v1.1.0
go: downloading github.com/pkg/browser v0.0.0-20240102092130-5ac0b6a4141c
go: downloading github.com/tklauser/numcpus v0.6.1
go: downloading go.opencensus.io v0.23.1-0.20220331163232-052120675fac
go: downloading go.opentelemetry.io/otel v1.24.0
go: downloading go.opentelemetry.io/otel/trace v1.24.0
go: downloading github.com/apache/thrift v0.16.0
go: downloading github.com/lestrrat-go/httprc v1.0.5
go: downloading github.com/lestrrat-go/blackmagic v1.0.2
go: downloading github.com/lestrrat-go/iter v1.0.2
go: downloading github.com/lestrrat-go/option v1.0.1
go: downloading github.com/dustin/go-humanize v1.0.1
go: downloading github.com/golang/glog v1.2.0
go: downloading github.com/kr/text v0.2.0
go: downloading github.com/rogpeppe/go-internal v1.12.0
go: downloading github.com/golang-jwt/jwt/v5 v5.2.0
go: downloading github.com/rivo/uniseg v0.4.7
go: downloading github.com/lestrrat-go/httpcc v1.0.1
go: downloading github.com/ncw/directio v1.0.5
go: downloading github.com/coocood/bbloom v0.0.0-20190830030839-58deb6228d64
go: downloading github.com/coocood/rtutil v0.0.0-20190304133409-c84515f646f2
go: downloading github.com/klauspost/cpuid v1.3.1
go: downloading github.com/golang-jwt/jwt v3.2.2+incompatible
go: downloading github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da
go: downloading github.com/go-logr/logr v1.4.1
go: downloading go.opentelemetry.io/otel/metric v1.24.0
go: downloading github.com/go-logr/stdr v1.2.2
go: downloading github.com/DataDog/zstd v1.5.5
go: downloading github.com/cockroachdb/tokenbucket v0.0.0-20230807174530-cc333fc44b06
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/changefeed_error/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
start tidb cluster in /tmp/tidb_cdc_test/multi_capture
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
[Pipeline] }
Cache not saved (inner-step execution failed)
go: downloading github.com/jmespath/go-jmespath v0.4.0
go: downloading github.com/google/s2a-go v0.1.7
go: downloading go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp v0.49.0
go: downloading github.com/googleapis/enterprise-certificate-proxy v0.3.2
go: downloading go.opentelemetry.io/contrib/instrumentation/google.golang.org/grpc/otelgrpc v0.49.0
go: downloading github.com/felixge/httpsnoop v1.0.4
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
go: downloading github.com/modern-go/reflect2 v1.0.2
go: downloading github.com/json-iterator/go v1.1.12
go: downloading github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd
start tidb cluster in /tmp/tidb_cdc_test/changefeed_error
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
check diff failed 1-th time, retry later
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/mq_sink_dispatcher/run.sh: line 1: 14404 Killed                  cdc_kafka_consumer --upstream-uri $SINK_URI --downstream-uri="mysql://root@127.0.0.1:3306/?safe-mode=true&batch-dml-enable=false" --upstream-tidb-dsn="root@tcp(${UP_TIDB_HOST}:${UP_TIDB_PORT})/?" --config="$CUR/conf/new_changefeed.toml" 2>&1
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/kafka_column_selector/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
check diff failed 2-th time, retry later
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
table multi_tables_ddl_test.t55 exists
table multi_tables_ddl_test.t66 exists
table multi_tables_ddl_test.t7 exists
table multi_tables_ddl_test.t88 exists
table multi_tables_ddl_test.finish_mark not exists for 1-th check, retry later
TEST FAILED: OUTPUT DOES NOT CONTAIN 'id: 1'
____________________________________
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
check data failed 1-th time, retry later
check data successfully
wait process cdc.test exit for 1-th time...
wait process cdc.test exit for 2-th time...
cdc.test: no process found
wait process cdc.test exit for 3-th time...
process cdc.test already exit
[Sat May  4 19:44:29 CST 2024] <<<<<< run test case ddl_puller_lag success! >>>>>>
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dae20740014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs, pid:7452, start at 2024-05-04 19:44:31.806697398 +0800 CST m=+5.101398990	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:31.812 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:31.773 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:31.773 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dae20740014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs, pid:7452, start at 2024-05-04 19:44:31.806697398 +0800 CST m=+5.101398990	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:31.812 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:31.773 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:31.773 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dae22b40014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-bft7v-7bqfs, pid:7529, start at 2024-05-04 19:44:31.937889737 +0800 CST m=+5.173191576	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:31.944 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:31.917 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:31.917 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/multi_source/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/multi_source/tiflash/log/error.log
arg matches is ArgMatches { args: {"advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/multi_source/tiflash/log/proxy.log"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/multi_source/tiflash/db/proxy"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/multi_source/tiflash-proxy.toml"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
start tidb cluster in /tmp/tidb_cdc_test/kafka_column_selector
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
check diff failed 3-th time, retry later
table multi_tables_ddl_test.finish_mark exists
check table exists success
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=test-normal
+ expected_state=normal
+ error_msg=null
+ tls_dir=
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c test-normal -s
+ info='{
  "upstream_id": 7365108887230533730,
  "namespace": "default",
  "id": "test-normal",
  "state": "normal",
  "checkpoint_tso": 449530579085950980,
  "checkpoint_time": "2024-05-04 19:44:30.854",
  "error": null
}'
+ echo '{
  "upstream_id": 7365108887230533730,
  "namespace": "default",
  "id": "test-normal",
  "state": "normal",
  "checkpoint_tso": 449530579085950980,
  "checkpoint_time": "2024-05-04 19:44:30.854",
  "error": null
}'
{
  "upstream_id": 7365108887230533730,
  "namespace": "default",
  "id": "test-normal",
  "state": "normal",
  "checkpoint_tso": 449530579085950980,
  "checkpoint_time": "2024-05-04 19:44:30.854",
  "error": null
}
++ echo '{' '"upstream_id":' 7365108887230533730, '"namespace":' '"default",' '"id":' '"test-normal",' '"state":' '"normal",' '"checkpoint_tso":' 449530579085950980, '"checkpoint_time":' '"2024-05-04' '19:44:30.854",' '"error":' null '}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \n\o\r\m\a\l ]]
++ echo '{' '"upstream_id":' 7365108887230533730, '"namespace":' '"default",' '"id":' '"test-normal",' '"state":' '"normal",' '"checkpoint_tso":' 449530579085950980, '"checkpoint_time":' '"2024-05-04' '19:44:30.854",' '"error":' null '}'
++ jq -r .error.message
+ message=null
+ [[ ! null =~ null ]]
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=test-error-1
+ expected_state=normal
+ error_msg=null
+ tls_dir=
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c test-error-1 -s
+ info='{
  "upstream_id": 7365108887230533730,
  "namespace": "default",
  "id": "test-error-1",
  "state": "normal",
  "checkpoint_tso": 449530579728465923,
  "checkpoint_time": "2024-05-04 19:44:33.305",
  "error": null
}'
+ echo '{
  "upstream_id": 7365108887230533730,
  "namespace": "default",
  "id": "test-error-1",
  "state": "normal",
  "checkpoint_tso": 449530579728465923,
  "checkpoint_time": "2024-05-04 19:44:33.305",
  "error": null
}'
{
  "upstream_id": 7365108887230533730,
  "namespace": "default",
  "id": "test-error-1",
  "state": "normal",
  "checkpoint_tso": 449530579728465923,
  "checkpoint_time": "2024-05-04 19:44:33.305",
  "error": null
}
++ echo '{' '"upstream_id":' 7365108887230533730, '"namespace":' '"default",' '"id":' '"test-error-1",' '"state":' '"normal",' '"checkpoint_tso":' 449530579728465923, '"checkpoint_time":' '"2024-05-04' '19:44:33.305",' '"error":' null '}'
++ jq -r .state
+ state=normal
+ [[ ! normal == \n\o\r\m\a\l ]]
++ echo '{' '"upstream_id":' 7365108887230533730, '"namespace":' '"default",' '"id":' '"test-error-1",' '"state":' '"normal",' '"checkpoint_tso":' 449530579728465923, '"checkpoint_time":' '"2024-05-04' '19:44:33.305",' '"error":' null '}'
++ jq -r .error.message
+ message=null
+ [[ ! null =~ null ]]
+ endpoints=http://127.0.0.1:2379
+ changefeed_id=test-error-2
+ expected_state=failed
+ error_msg=ErrSyncRenameTableFailed
+ tls_dir=
+ [[ http://127.0.0.1:2379 =~ https ]]
++ cdc cli changefeed query --pd=http://127.0.0.1:2379 -c test-error-2 -s
+ info='{
  "upstream_id": 7365108887230533730,
  "namespace": "default",
  "id": "test-error-2",
  "state": "failed",
  "checkpoint_tso": 449530578942033926,
  "checkpoint_time": "2024-05-04 19:44:30.305",
  "error": {
    "time": "2024-05-04T19:44:31.39858409+08:00",
    "addr": "127.0.0.1:8300",
    "code": "CDC:ErrSyncRenameTableFailed",
    "message": "[CDC:ErrSyncRenameTableFailed]table'\''s old name is not in filter rule, and its new name in filter rule table id '\''130'\'', ddl query: [rename table t11 to t9], it'\''s an unexpected behavior, if you want to replicate this table, please add its old name to filter rule."
  }
}'
+ echo '{
  "upstream_id": 7365108887230533730,
  "namespace": "default",
  "id": "test-error-2",
  "state": "failed",
  "checkpoint_tso": 449530578942033926,
  "checkpoint_time": "2024-05-04 19:44:30.305",
  "error": {
    "time": "2024-05-04T19:44:31.39858409+08:00",
    "addr": "127.0.0.1:8300",
    "code": "CDC:ErrSyncRenameTableFailed",
    "message": "[CDC:ErrSyncRenameTableFailed]table'\''s old name is not in filter rule, and its new name in filter rule table id '\''130'\'', ddl query: [rename table t11 to t9], it'\''s an unexpected behavior, if you want to replicate this table, please add its old name to filter rule."
  }
}'
{
  "upstream_id": 7365108887230533730,
  "namespace": "default",
  "id": "test-error-2",
  "state": "failed",
  "checkpoint_tso": 449530578942033926,
  "checkpoint_time": "2024-05-04 19:44:30.305",
  "error": {
    "time": "2024-05-04T19:44:31.39858409+08:00",
    "addr": "127.0.0.1:8300",
    "code": "CDC:ErrSyncRenameTableFailed",
    "message": "[CDC:ErrSyncRenameTableFailed]table's old name is not in filter rule, and its new name in filter rule table id '130', ddl query: [rename table t11 to t9], it's an unexpected behavior, if you want to replicate this table, please add its old name to filter rule."
  }
}
++ jq -r .state
++ echo '{' '"upstream_id":' 7365108887230533730, '"namespace":' '"default",' '"id":' '"test-error-2",' '"state":' '"failed",' '"checkpoint_tso":' 449530578942033926, '"checkpoint_time":' '"2024-05-04' '19:44:30.305",' '"error":' '{' '"time":' '"2024-05-04T19:44:31.39858409+08:00",' '"addr":' '"127.0.0.1:8300",' '"code":' '"CDC:ErrSyncRenameTableFailed",' '"message":' '"[CDC:ErrSyncRenameTableFailed]table'\''s' old name is not in filter rule, and its new name in filter rule table id ''\''130'\'',' ddl query: '[rename' table t11 to 't9],' 'it'\''s' an unexpected behavior, if you want to replicate this table, please add its old name to filter 'rule."' '}' '}'
+ state=failed
+ [[ ! failed == \f\a\i\l\e\d ]]
++ jq -r .error.message
++ echo '{' '"upstream_id":' 7365108887230533730, '"namespace":' '"default",' '"id":' '"test-error-2",' '"state":' '"failed",' '"checkpoint_tso":' 449530578942033926, '"checkpoint_time":' '"2024-05-04' '19:44:30.305",' '"error":' '{' '"time":' '"2024-05-04T19:44:31.39858409+08:00",' '"addr":' '"127.0.0.1:8300",' '"code":' '"CDC:ErrSyncRenameTableFailed",' '"message":' '"[CDC:ErrSyncRenameTableFailed]table'\''s' old name is not in filter rule, and its new name in filter rule table id ''\''130'\'',' ddl query: '[rename' table t11 to 't9],' 'it'\''s' an unexpected behavior, if you want to replicate this table, please add its old name to filter 'rule."' '}' '}'
+ message='[CDC:ErrSyncRenameTableFailed]table'\''s old name is not in filter rule, and its new name in filter rule table id '\''130'\'', ddl query: [rename table t11 to t9], it'\''s an unexpected behavior, if you want to replicate this table, please add its old name to filter rule.'
+ [[ ! [CDC:ErrSyncRenameTableFailed]table's old name is not in filter rule, and its new name in filter rule table id '130', ddl query: [rename table t11 to t9], it's an unexpected behavior, if you want to replicate this table, please add its old name to filter rule. =~ ErrSyncRenameTableFailed ]]
check diff successfully
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.multi_source.cli.8850.out cli tso query --pd=http://127.0.0.1:2379
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
wait process cdc.test exit for 1-th time...
check diff failed 4-th time, retry later
[2024/05/04 19:44:26.370 +08:00] [INFO] [main.go:99] ["running ddl test: 1 modifyColumnDefaultValueDDL2"]
[2024/05/04 19:44:26.370 +08:00] [INFO] [main.go:99] ["running ddl test: 0 modifyColumnDefaultValueDDL1"]
[2024/05/04 19:44:26.459 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs31d46f2f_4bae_42d7_96fb_8ebda86f474d"]
[2024/05/04 19:44:26.488 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs5fe8d828_e9e9_47b7_b14b_1564de901c65"]
[2024/05/04 19:44:26.602 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLsa06aeca8_91c3_409c_8022_7dfb130595d8"]
[2024/05/04 19:44:26.675 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs34776f4a_0a14_4f6a_bc0c_9f35d954ea71"]
[2024/05/04 19:44:26.683 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs584d0eb7_6a54_49ec_9318_aceb7c9d06c6"]
[2024/05/04 19:44:26.691 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs35095ce3_e5f7_4f32_ac8c_0bfddd1dab79"]
[2024/05/04 19:44:26.760 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:26.791 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:26.795 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:26.845 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:26.896 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs85f1c604_77c7_492b_a8c4_b5cf49257676"]
[2024/05/04 19:44:26.898 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLseeb608f8_ce6f_4ee1_a1e9_211c7b8bf7d8"]
[2024/05/04 19:44:26.962 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:26.964 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:27.052 +08:00] [INFO] [main.go:178] ["1 insert success: 100"]
[2024/05/04 19:44:27.068 +08:00] [INFO] [main.go:178] ["1 insert success: 100"]
[2024/05/04 19:44:27.157 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:27.167 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:27.176 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:27.177 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:27.251 +08:00] [INFO] [main.go:178] ["0 insert success: 100"]
[2024/05/04 19:44:27.268 +08:00] [INFO] [main.go:178] ["0 insert success: 100"]
[2024/05/04 19:44:27.272 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:27.280 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:27.289 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:27.365 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:27.368 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:27.379 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:27.480 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:27.482 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:27.483 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:27.550 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:27.551 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:27.566 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:27.595 +08:00] [INFO] [main.go:178] ["1 insert success: 200"]
[2024/05/04 19:44:27.646 +08:00] [INFO] [main.go:178] ["1 insert success: 200"]
[2024/05/04 19:44:27.688 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:27.756 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:27.767 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:27.767 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:27.871 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:27.882 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:27.887 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:27.977 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:27.990 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:28.055 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:28.061 +08:00] [INFO] [main.go:178] ["0 insert success: 200"]
[2024/05/04 19:44:28.067 +08:00] [INFO] [main.go:199] ["0 delete success: 100"]
[2024/05/04 19:44:28.080 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:28.102 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:28.144 +08:00] [INFO] [main.go:178] ["0 insert success: 200"]
[2024/05/04 19:44:28.150 +08:00] [INFO] [main.go:199] ["0 delete success: 100"]
[2024/05/04 19:44:28.153 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:28.157 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:28.159 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:28.178 +08:00] [INFO] [main.go:178] ["1 insert success: 300"]
[2024/05/04 19:44:28.178 +08:00] [INFO] [main.go:178] ["1 insert success: 300"]
[2024/05/04 19:44:28.183 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:28.258 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:28.292 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:28.385 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:28.449 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:28.449 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:28.460 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:28.549 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:28.566 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:28.660 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:28.669 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:28.680 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:28.686 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:28.750 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:28.756 +08:00] [INFO] [main.go:178] ["1 insert success: 400"]
[2024/05/04 19:44:28.758 +08:00] [INFO] [main.go:178] ["1 insert success: 400"]
[2024/05/04 19:44:28.770 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:28.772 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:28.789 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:28.850 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:28.881 +08:00] [INFO] [main.go:178] ["0 insert success: 300"]
[2024/05/04 19:44:28.882 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:28.972 +08:00] [INFO] [main.go:178] ["0 insert success: 300"]
[2024/05/04 19:44:28.976 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:28.985 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:29.045 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:29.050 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:29.090 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:29.165 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:29.252 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:29.253 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:29.277 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:29.282 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:29.352 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:29.353 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:29.357 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:29.366 +08:00] [INFO] [main.go:178] ["1 insert success: 500"]
[2024/05/04 19:44:29.367 +08:00] [INFO] [main.go:178] ["1 insert success: 500"]
[2024/05/04 19:44:29.375 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:29.455 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:29.554 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:29.577 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:29.657 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:29.687 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:29.689 +08:00] [INFO] [main.go:178] ["0 insert success: 400"]
[2024/05/04 19:44:29.747 +08:00] [INFO] [main.go:199] ["0 delete success: 200"]
[2024/05/04 19:44:29.754 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:29.756 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:29.792 +08:00] [INFO] [main.go:178] ["0 insert success: 400"]
[2024/05/04 19:44:29.848 +08:00] [INFO] [main.go:199] ["0 delete success: 200"]
[2024/05/04 19:44:29.868 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:29.874 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:29.890 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:29.960 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:29.964 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:29.966 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:29.970 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:29.980 +08:00] [INFO] [main.go:178] ["1 insert success: 600"]
[2024/05/04 19:44:29.982 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:29.986 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:29.993 +08:00] [INFO] [main.go:178] ["1 insert success: 600"]
[2024/05/04 19:44:30.063 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:30.147 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:30.164 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:30.251 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:30.272 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:30.287 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:30.345 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:30.459 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:30.468 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:30.474 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:30.488 +08:00] [INFO] [main.go:178] ["0 insert success: 500"]
[2024/05/04 19:44:30.492 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:30.548 +08:00] [INFO] [main.go:178] ["1 insert success: 700"]
[2024/05/04 19:44:30.562 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:30.562 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:30.571 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:30.575 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:30.575 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:30.576 +08:00] [INFO] [main.go:178] ["1 insert success: 700"]
[2024/05/04 19:44:30.587 +08:00] [INFO] [main.go:178] ["0 insert success: 500"]
[2024/05/04 19:44:30.650 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:30.673 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:30.693 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:30.778 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:30.848 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:30.872 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:30.882 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:31.047 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:31.060 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:31.072 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:31.085 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:31.086 +08:00] [INFO] [main.go:178] ["1 insert success: 800"]
[2024/05/04 19:44:31.150 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:31.150 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:31.209 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:31.219 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:31.224 +08:00] [INFO] [main.go:178] ["1 insert success: 800"]
[2024/05/04 19:44:31.233 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:31.253 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:31.271 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:31.286 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:31.288 +08:00] [INFO] [main.go:178] ["0 insert success: 600"]
[2024/05/04 19:44:31.292 +08:00] [INFO] [main.go:199] ["0 delete success: 300"]
[2024/05/04 19:44:31.361 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:31.384 +08:00] [INFO] [main.go:178] ["0 insert success: 600"]
[2024/05/04 19:44:31.389 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:31.389 +08:00] [INFO] [main.go:199] ["0 delete success: 300"]
[2024/05/04 19:44:31.474 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:31.480 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:31.655 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:31.657 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:31.667 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:31.675 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:31.685 +08:00] [INFO] [main.go:178] ["1 insert success: 900"]
[2024/05/04 19:44:31.748 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:31.752 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:31.848 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:31.849 +08:00] [INFO] [main.go:178] ["1 insert success: 900"]
[2024/05/04 19:44:31.852 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:31.857 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:31.871 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:31.877 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:31.947 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:31.960 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:31.986 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:32.072 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:32.081 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:32.084 +08:00] [INFO] [main.go:178] ["0 insert success: 700"]
[2024/05/04 19:44:32.188 +08:00] [INFO] [main.go:178] ["0 insert success: 700"]
[2024/05/04 19:44:32.245 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:32.253 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:32.266 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:32.275 +08:00] [INFO] [main.go:178] ["1 insert success: 1000"]
[2024/05/04 19:44:32.291 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:32.346 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:32.346 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:32.391 +08:00] [INFO] [main.go:178] ["1 insert success: 1000"]
[2024/05/04 19:44:32.447 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:32.455 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:32.460 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:32.465 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:32.495 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:32.547 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:32.557 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:32.574 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:32.659 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:32.666 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:32.783 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:32.848 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:32.863 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:32.872 +08:00] [INFO] [main.go:178] ["1 insert success: 1100"]
[2024/05/04 19:44:32.946 +08:00] [INFO] [main.go:178] ["0 insert success: 800"]
[2024/05/04 19:44:32.947 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:32.948 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:32.952 +08:00] [INFO] [main.go:199] ["0 delete success: 400"]
[2024/05/04 19:44:32.966 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:32.986 +08:00] [INFO] [main.go:178] ["1 insert success: 1100"]
[2024/05/04 19:44:33.046 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:33.059 +08:00] [INFO] [main.go:178] ["0 insert success: 800"]
[2024/05/04 19:44:33.063 +08:00] [INFO] [main.go:199] ["0 delete success: 400"]
[2024/05/04 19:44:33.065 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:33.065 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:33.070 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:33.148 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:33.171 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:33.172 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:33.179 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:33.253 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:33.264 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:33.367 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:33.380 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:33.454 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:33.472 +08:00] [INFO] [main.go:178] ["1 insert success: 1200"]
[2024/05/04 19:44:33.482 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:33.486 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:33.576 +08:00] [INFO] [main.go:178] ["1 insert success: 1200"]
[2024/05/04 19:44:33.648 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:33.655 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:33.676 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:33.679 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:33.679 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:33.766 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:33.778 +08:00] [INFO] [main.go:178] ["0 insert success: 900"]
[2024/05/04 19:44:33.793 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:33.850 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:33.866 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:33.871 +08:00] [INFO] [main.go:178] ["0 insert success: 900"]
[2024/05/04 19:44:33.873 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:33.886 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:33.978 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:34.045 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:34.054 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:34.080 +08:00] [INFO] [main.go:178] ["1 insert success: 1300"]
[2024/05/04 19:44:34.084 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:34.086 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:34.166 +08:00] [INFO] [main.go:178] ["1 insert success: 1300"]
[2024/05/04 19:44:34.245 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:34.252 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:34.270 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:34.274 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:34.275 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:34.350 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:34.377 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:34.381 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:34.458 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:34.465 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:34.470 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:34.559 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:34.566 +08:00] [INFO] [main.go:178] ["0 insert success: 1000"]
[2024/05/04 19:44:34.571 +08:00] [INFO] [main.go:199] ["0 delete success: 500"]
[2024/05/04 19:44:34.576 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:34.584 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:34.657 +08:00] [INFO] [main.go:178] ["1 insert success: 1400"]
[2024/05/04 19:44:34.666 +08:00] [INFO] [main.go:178] ["0 insert success: 1000"]
[2024/05/04 19:44:34.670 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:34.671 +08:00] [INFO] [main.go:199] ["0 delete success: 500"]
[2024/05/04 19:44:34.682 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:34.755 +08:00] [INFO] [main.go:178] ["1 insert success: 1400"]
[2024/05/04 19:44:34.845 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:34.850 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:34.865 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:34.866 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:34.866 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:34.894 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:34.963 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:34.974 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:34.994 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:35.059 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:35.069 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:35.150 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:35.163 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:35.167 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:35.251 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:35.253 +08:00] [INFO] [main.go:178] ["1 insert success: 1500"]
[2024/05/04 19:44:35.264 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:35.275 +08:00] [INFO] [main.go:178] ["1 insert success: 1500"]
[2024/05/04 19:44:35.373 +08:00] [INFO] [main.go:178] ["0 insert success: 1100"]
[2024/05/04 19:44:35.379 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:35.449 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:35.453 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:35.454 +08:00] [INFO] [main.go:178] ["0 insert success: 1100"]
[2024/05/04 19:44:35.464 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:35.473 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:35.492 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:35.549 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:35.569 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:35.583 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:35.661 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:35.760 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:35.776 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:35.777 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:35.854 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:35.870 +08:00] [INFO] [main.go:178] ["1 insert success: 1600"]
[2024/05/04 19:44:35.873 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:35.885 +08:00] [INFO] [main.go:178] ["1 insert success: 1600"]
wait process cdc.test exit for 2-th time...
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // container
[2024/05/04 19:44:35.948 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:35.989 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:36.070 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:36.074 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:36.161 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:36.164 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:36.172 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:36.179 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[Pipeline] }
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
[2024/05/04 19:44:36.252 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:36.256 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:36.259 +08:00] [INFO] [main.go:178] ["0 insert success: 1200"]
[2024/05/04 19:44:36.263 +08:00] [INFO] [main.go:199] ["0 delete success: 600"]
[2024/05/04 19:44:36.274 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:36.364 +08:00] [INFO] [main.go:178] ["0 insert success: 1200"]
[2024/05/04 19:44:36.368 +08:00] [INFO] [main.go:199] ["0 delete success: 600"]
[2024/05/04 19:44:36.381 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:36.394 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
wait process cdc.test exit for 3-th time...
[Pipeline] // cache
[2024/05/04 19:44:36.455 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:36.473 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:36.477 +08:00] [INFO] [main.go:178] ["1 insert success: 1700"]
[2024/05/04 19:44:36.485 +08:00] [INFO] [main.go:178] ["1 insert success: 1700"]
[2024/05/04 19:44:36.498 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:36.572 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:36.598 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:36.648 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:36.685 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
+ set +x
+ tso='449530580257472513
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530580257472513 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
[Sat May  4 19:44:36 CST 2024] <<<<<< START cdc server in multi_source case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ GO_FAILPOINTS=
+ (( i = 0 ))
+ (( i <= 50 ))
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.multi_source.88868888.out server --log-file /tmp/tidb_cdc_test/multi_source/cdc.log --log-level debug --data-dir /tmp/tidb_cdc_test/multi_source/cdc_data --cluster-id default
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
[2024/05/04 19:44:36.749 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:36.750 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:36.754 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:36.761 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:36.772 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:36.883 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
[2024/05/04 19:44:36.948 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
cdc.test: no process found
wait process cdc.test exit for 4-th time...
process cdc.test already exit
[Sat May  4 19:44:36 CST 2024] <<<<<< run test case multi_tables_ddl_v2 success! >>>>>>
[2024/05/04 19:44:36.976 +08:00] [INFO] [main.go:178] ["1 insert success: 1800"]
[2024/05/04 19:44:36.986 +08:00] [INFO] [main.go:178] ["0 insert success: 1300"]
[2024/05/04 19:44:36.989 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:36.990 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:36.994 +08:00] [INFO] [main.go:178] ["1 insert success: 1800"]
[2024/05/04 19:44:37.057 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:37.061 +08:00] [INFO] [main.go:178] ["0 insert success: 1300"]
[2024/05/04 19:44:37.086 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:37.156 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
[2024/05/04 19:44:37.160 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
[2024/05/04 19:44:37.250 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:37.268 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:37.269 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
[2024/05/04 19:44:37.274 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:37.290 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:37.295 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
[2024/05/04 19:44:37.446 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[2024/05/04 19:44:37.454 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
[Pipeline] }
check diff failed 5-th time, retry later
[Pipeline] // timeout
[Pipeline] // timeout
[2024/05/04 19:44:37.485 +08:00] [INFO] [main.go:178] ["1 insert success: 1900"]
[2024/05/04 19:44:37.551 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:37.551 +08:00] [INFO] [main.go:178] ["1 insert success: 1900"]
[2024/05/04 19:44:37.577 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:37.593 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
[2024/05/04 19:44:37.668 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[2024/05/04 19:44:37.671 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[2024/05/04 19:44:37.746 +08:00] [INFO] [main.go:178] ["0 insert success: 1400"]
[2024/05/04 19:44:37.751 +08:00] [INFO] [main.go:199] ["0 delete success: 700"]
[2024/05/04 19:44:37.759 +08:00] [INFO] [main.go:178] ["0 insert success: 1400"]
[2024/05/04 19:44:37.759 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
[2024/05/04 19:44:37.763 +08:00] [INFO] [main.go:199] ["0 delete success: 700"]
[2024/05/04 19:44:37.777 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
[2024/05/04 19:44:37.796 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
[2024/05/04 19:44:37.891 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[2024/05/04 19:44:37.965 +08:00] [INFO] [main.go:178] ["1 insert success: 2000"]
[Pipeline] // withEnv
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
[Pipeline] }
[2024/05/04 19:44:37.971 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
[2024/05/04 19:44:38.001 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
[2024/05/04 19:44:38.005 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[2024/05/04 19:44:38.042 +08:00] [INFO] [main.go:178] ["73 insert success: 2000"]
[2024/05/04 19:44:38.169 +08:00] [INFO] [main.go:178] ["0 insert success: 1500"]
[2024/05/04 19:44:38.188 +08:00] [INFO] [main.go:178] ["73 insert success: 2000"]
[2024/05/04 19:44:38.255 +08:00] [INFO] [main.go:178] ["1 insert success: 2100"]
[2024/05/04 19:44:38.260 +08:00] [INFO] [main.go:178] ["72 insert success: 2000"]
[2024/05/04 19:44:38.291 +08:00] [INFO] [main.go:178] ["72 insert success: 2000"]
[2024/05/04 19:44:38.294 +08:00] [INFO] [main.go:178] ["73 insert success: 2000"]
[2024/05/04 19:44:38.400 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs3605734f_5815_46e3_92c3_e1eca893b942"]
[2024/05/04 19:44:38.473 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs92a216b9_9e9a_43de_b8c9_dfe698065d2f"]
[Pipeline] // dir
Post stage
[Pipeline] // container
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dae7b0c0006	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf, pid:7694, start at 2024-05-04 19:44:37.581100313 +0800 CST m=+5.324044029	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:37.588 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:37.571 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:37.571 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
[Pipeline] // container
[2024/05/04 19:44:38.476 +08:00] [INFO] [main.go:178] ["73 insert success: 2100"]
[2024/05/04 19:44:38.538 +08:00] [INFO] [main.go:178] ["0 insert success: 1600"]
[2024/05/04 19:44:38.541 +08:00] [INFO] [main.go:199] ["0 delete success: 800"]
[2024/05/04 19:44:38.543 +08:00] [INFO] [main.go:178] ["1 insert success: 2200"]
[2024/05/04 19:44:38.559 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLsee1567a1_8909_4368_8152_d6fa9aa335b6"]
[2024/05/04 19:44:38.660 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs0acfd437_8b6a_433f_9356_7e159db4f796"]
[2024/05/04 19:44:38.675 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs85359766_9a51_41b3_a8a3_6656160a5e9e"]
[2024/05/04 19:44:38.688 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:38.694 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[2024/05/04 19:44:38.770 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:38.774 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:38.834 +08:00] [INFO] [main.go:178] ["1 insert success: 2300"]
[2024/05/04 19:44:38.860 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:38.865 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:38.954 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLsf521a9a7_f09d_461a_987d_127efed8bed0"]
[2024/05/04 19:44:38.963 +08:00] [INFO] [main.go:178] ["0 insert success: 1700"]
[2024/05/04 19:44:38.971 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLsd67e1287_f1c2_43bb_852b_dec28cd8b2d7"]
[Pipeline] sh
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
[2024/05/04 19:44:38.999 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:39.045 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:39.060 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:39.069 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:39.071 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:39.090 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs69587118_ced4_43e7_824c_43ed80c751fe"]
[2024/05/04 19:44:39.114 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:39.174 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:39.179 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[Pipeline] sh
[2024/05/04 19:44:39.252 +08:00] [INFO] [main.go:178] ["1 insert success: 2400"]
[2024/05/04 19:44:39.268 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:39.282 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:39.445 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:39.449 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:39.455 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:39.465 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:39.487 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G13
Run cases: tiflash region_merge common_1
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=e148f589-f8ed-445b-8978-ebed37e7a973
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-38244-hqb2r
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G13
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-38244
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-38244 pingcap-tiflow-pull-cdc-integration-kafka-test-1845-38244-hqb2r
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-38244-hqb2r
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/tiflash/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
[Pipeline] sh
[2024/05/04 19:44:39.554 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:39.570 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:39.578 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:39.583 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:39.594 +08:00] [INFO] [main.go:178] ["0 insert success: 1800"]
[2024/05/04 19:44:39.597 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:39.646 +08:00] [INFO] [main.go:199] ["0 delete success: 900"]
[2024/05/04 19:44:39.654 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:39.663 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:39.690 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:39.694 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[Pipeline] sh
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
[2024/05/04 19:44:39.779 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:39.788 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:39.889 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:39.896 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:39.898 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:39.950 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:39.970 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:39.990 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:40.005 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
check diff successfully
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G12
Run cases: many_pk_or_uk capture_session_done_during_task ddl_attributes
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=0779420d-d256-4324-86a8-d4712362e311
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-tgsj5-209qp
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G12
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-tgsj5
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-tgsj5-209qp pingcap_tiflow_pull_cdc_integration_kafka_test_1845-tgsj5
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-tgsj5-209qp
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/many_pk_or_uk/run.sh using Sink-Type: kafka... <<=================
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Sat, 04 May 2024 11:44:39 GMT
< Content-Length: 815
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/58d75a3e-4b5e-417b-ae74-63c72ec94aca
	{"id":"58d75a3e-4b5e-417b-ae74-63c72ec94aca","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823077}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436b6234d2
	58d75a3e-4b5e-417b-ae74-63c72ec94aca

/tidb/cdc/default/default/upstream/7365108970966329384
	{"id":7365108970966329384,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/58d75a3e-4b5e-417b-ae74-63c72ec94aca
	{"id":"58d75a3e-4b5e-417b-ae74-63c72ec94aca","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823077}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436b6234d2
	58d75a3e-4b5e-417b-ae74-63c72ec94aca

/tidb/cdc/default/default/upstream/7365108970966329384
	{"id":7365108970966329384,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/58d75a3e-4b5e-417b-ae74-63c72ec94aca
	{"id":"58d75a3e-4b5e-417b-ae74-63c72ec94aca","address":"127.0.0.1:8300","version":"v8.2.0-alpha-88-g57a3c8e82","git-hash":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/bin/cdc.test","start-timestamp":1714823077}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f436b6234d2
	58d75a3e-4b5e-417b-ae74-63c72ec94aca

/tidb/cdc/default/default/upstream/7365108970966329384
	{"id":7365108970966329384,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.multi_source.cli.8947.out cli changefeed create --start-ts=449530580257472513 '--sink-uri=kafka://127.0.0.1:9092/ticdc-multi-source-test-16481?protocol=open-protocol&partition-num=4&kafka-version=2.4.1&max-message-bytes=10485760'
[Pipeline] sh
[2024/05/04 19:44:40.046 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:40.050 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:40.071 +08:00] [INFO] [main.go:178] ["72 insert success: 200"]
[2024/05/04 19:44:40.080 +08:00] [INFO] [main.go:178] ["73 insert success: 200"]
[2024/05/04 19:44:40.087 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:40.145 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:40.149 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:40.189 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:40.198 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
Create changefeed successfully!
ID: ebe3bdaf-93a1-4e68-bb81-d44e2f405b24
Info: {"upstream_id":7365108970966329384,"namespace":"default","id":"ebe3bdaf-93a1-4e68-bb81-d44e2f405b24","sink_uri":"kafka://127.0.0.1:9092/ticdc-multi-source-test-16481?protocol=open-protocol\u0026partition-num=4\u0026kafka-version=2.4.1\u0026max-message-bytes=10485760","create_time":"2024-05-04T19:44:40.330757588+08:00","start_ts":449530580257472513,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"protocol":"open-protocol","csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.2.0-alpha-88-g57a3c8e82","resolved_ts":449530580257472513,"checkpoint_ts":449530580257472513,"checkpoint_time":"2024-05-04 19:44:35.323"}
PASS
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G14
Run cases: changefeed_finish force_replicate_table
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=d4912dfd-3331-4064-a0ca-e17b7730e1c1
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d59wt-f77sf
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G14
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-d59wt
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d59wt-f77sf pingcap_tiflow_pull_cdc_integration_kafka_test_1845-d59wt
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d59wt-f77sf
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/changefeed_finish/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
[Pipeline] sh
[2024/05/04 19:44:40.292 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:40.294 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:40.300 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:40.359 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:40.380 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:40.398 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:40.447 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:40.454 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:40.459 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:40.480 +08:00] [INFO] [main.go:178] ["72 insert success: 300"]
[2024/05/04 19:44:40.490 +08:00] [INFO] [main.go:178] ["73 insert success: 300"]
[2024/05/04 19:44:40.497 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dae7b0c0006	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf, pid:7694, start at 2024-05-04 19:44:37.581100313 +0800 CST m=+5.324044029	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:37.588 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:37.571 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:37.571 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dae7bc00015	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-t0ptc-wtbgf, pid:7773, start at 2024-05-04 19:44:37.651760809 +0800 CST m=+5.345951622	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:37.658 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:37.616 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:37.616 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/multi_capture/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/multi_capture/tiflash/log/error.log
arg matches is ArgMatches { args: {"config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/multi_capture/tiflash-proxy.toml"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/multi_capture/tiflash/db/proxy"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/multi_capture/tiflash/log/proxy.log"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
coverage: 2.4% of statements in github.com/pingcap/tiflow/...
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G15
Run cases: new_ci_collation batch_add_table multi_rocks
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=c3df7314-5277-4d1f-80b1-2feca7473b34
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-s7s3l-7txfz
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G15
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-s7s3l
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-s7s3l pingcap-tiflow-pull-cdc-integration-kafka-test-1845-s7s3l-7txfz
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-s7s3l-7txfz
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/new_ci_collation/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
[2024/05/04 19:44:40.551 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:40.552 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:40.593 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:40.595 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:40.672 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:40.689 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:40.692 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:40.747 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[Pipeline] // node
[Pipeline] }
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G16
Run cases: owner_resign processor_etcd_worker_delay sink_hang
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=91b3b8db-095f-46a2-a663-42031ca44b35
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6gzn1-2sfnd
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G16
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-6gzn1
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-6gzn1 pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6gzn1-2sfnd
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-6gzn1-2sfnd
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/owner_resign/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:44:40 CST 2024] <<<<<< run test case owner_resign success! >>>>>>
[Pipeline] sh
[2024/05/04 19:44:40.773 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:40.797 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:40.865 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:40.865 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:40.873 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:40.895 +08:00] [INFO] [main.go:178] ["72 insert success: 400"]
[2024/05/04 19:44:40.904 +08:00] [INFO] [main.go:178] ["73 insert success: 400"]
[2024/05/04 19:44:40.908 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:40.957 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:40.966 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:40.996 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:41.004 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:41.076 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:41.089 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:41.099 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:41.148 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:41.176 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:41.247 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
+ rm -rf /tmp/tidb_cdc_test
+ mkdir -p /tmp/tidb_cdc_test
+ chmod +x ./tests/integration_tests/run_group.sh
+ ./tests/integration_tests/run_group.sh kafka G17
+ ls /tmp/tidb_cdc_test/
cov.row_format.cli.2837.out
cov.row_format.cli.2934.out
row_format
sql_res.row_format.txt
++ find /tmp/tidb_cdc_test/ -type f -name '*.log'
+ tar -cvzf log-G03.tar.gz /tmp/tidb_cdc_test/row_format/tikv3/db/000005.log /tmp/tidb_cdc_test/row_format/tikv_down.log /tmp/tidb_cdc_test/row_format/down_pd.log /tmp/tidb_cdc_test/row_format/tikv2/db/000005.log /tmp/tidb_cdc_test/row_format/tikv_down/db/000005.log /tmp/tidb_cdc_test/row_format/cdc.log /tmp/tidb_cdc_test/row_format/stdout.log /tmp/tidb_cdc_test/row_format/tidb-slow.log /tmp/tidb_cdc_test/row_format/tikv1.log /tmp/tidb_cdc_test/row_format/tidb.log /tmp/tidb_cdc_test/row_format/tiflash/db/proxy/db/000005.log /tmp/tidb_cdc_test/row_format/tiflash/log/error.log /tmp/tidb_cdc_test/row_format/tiflash/log/server.log /tmp/tidb_cdc_test/row_format/tiflash/log/proxy.log /tmp/tidb_cdc_test/row_format/tidb_down.log /tmp/tidb_cdc_test/row_format/tikv1/db/000005.log /tmp/tidb_cdc_test/row_format/pd1.log /tmp/tidb_cdc_test/row_format/tidb_other.log /tmp/tidb_cdc_test/row_format/down_pd/hot-region/000001.log /tmp/tidb_cdc_test/row_format/down_pd/region-meta/000001.log /tmp/tidb_cdc_test/row_format/tikv3.log /tmp/tidb_cdc_test/row_format/cdc_kafka_consumer.log /tmp/tidb_cdc_test/row_format/pd1/hot-region/000001.log /tmp/tidb_cdc_test/row_format/pd1/region-meta/000001.log /tmp/tidb_cdc_test/row_format/tikv2.log /tmp/tidb_cdc_test/row_format/cdc_kafka_consumer_stdout.log /tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0004/000002.log /tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0003/000002.log /tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0001/000002.log /tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0005/000002.log /tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0007/000002.log /tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0006/000002.log /tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0000/000002.log /tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0002/000002.log
tar: Removing leading `/' from member names
/tmp/tidb_cdc_test/row_format/tikv3/db/000005.log
/tmp/tidb_cdc_test/row_format/tikv_down.log
[2024/05/04 19:44:41.288 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:41.288 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:41.298 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:41.366 +08:00] [INFO] [main.go:178] ["72 insert success: 500"]
[2024/05/04 19:44:41.376 +08:00] [INFO] [main.go:178] ["73 insert success: 500"]
[2024/05/04 19:44:41.379 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:41.400 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:41.452 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:41.476 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:41.483 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
Run cases: clustered_index processor_resolved_ts_fallback
PROW_JOB_ID=02167274-41fb-4ee5-bed5-8bce85c77132
JENKINS_NODE_COOKIE=df335106-c3a2-4885-9632-b83ee028180f
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d1pdj-fsd6q
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT=tcp://10.233.0.1:443
KUBERNETES_PORT_443_TCP_PORT=443
TERM=xterm
STAGE_NAME=Test
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786719159128690688","prowjobid":"02167274-41fb-4ee5-bed5-8bce85c77132","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/57a3c8e82a4bc7fa16bc398b80fe6e8f0f61cfce","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/_utils:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../bin:/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/../../scripts/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
HUDSON_URL=https://do.pingcap.net/jenkins/
TICDC_COVERALLS_TOKEN=****
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1845
TEST_GROUP=G17
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786719159128690688
TICDC_CODECOV_TOKEN=****
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1845/display/redirect?page=tests
SHLVL=5
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-d1pdj
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-8f1d433d6527e85b4c28b432b07fa0c56dd3090c9176381d29a5b4531676247a
NODE_LABELS=pingcap_tiflow_pull_cdc_integration_kafka_test_1845-d1pdj pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d1pdj-fsd6q
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1845-d1pdj-fsd6q
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1845
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
_=/usr/bin/env
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/clustered_index/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:44:41 CST 2024] <<<<<< skip test case clustered_index for kafka! >>>>>>
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/processor_resolved_ts_fallback/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:44:41 CST 2024] <<<<<< run test case processor_resolved_ts_fallback success! >>>>>>
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
/tmp/tidb_cdc_test/row_format/down_pd.log
/tmp/tidb_cdc_test/row_format/tikv2/db/000005.log
/tmp/tidb_cdc_test/row_format/tikv_down/db/000005.log
/tmp/tidb_cdc_test/row_format/cdc.log
[2024/05/04 19:44:41.554 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:41.566 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:41.571 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
[2024/05/04 19:44:41.582 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:41.648 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:41.668 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:41.750 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:41.751 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:41.766 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:41.786 +08:00] [INFO] [main.go:178] ["72 insert success: 600"]
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dae8f6c0018	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd, pid:4291, start at 2024-05-04 19:44:38.909304498 +0800 CST m=+5.240231273	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:38.916 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:38.875 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:38.875 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dae8f6c0018	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd, pid:4291, start at 2024-05-04 19:44:38.909304498 +0800 CST m=+5.240231273	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:38.916 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:38.875 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:38.875 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0dae90f80014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-321md-f5hpd, pid:4369, start at 2024-05-04 19:44:39.007932317 +0800 CST m=+5.288571965	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:39.015 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:38.974 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:38.974 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
[Pipeline] // podTemplate
[Pipeline] }
+ set +x
[Sat May  4 19:44:41 CST 2024] <<<<<< START kafka consumer in multi_source case >>>>>>
go: downloading github.com/pingcap/log v1.1.1-0.20240314023424-862ccc32f18d
go: downloading github.com/pingcap/errors v0.11.5-0.20240318064555-6bd07397691f
go: downloading go.uber.org/zap v1.27.0
go: downloading github.com/pingcap/tidb-tools v0.0.0-20240305021104-9f9bea84490b
go: downloading github.com/BurntSushi/toml v1.3.2
go: downloading github.com/pingcap/tidb v1.1.0-beta.0.20240415145106-cd9c676e9ba4
go: downloading gopkg.in/natefinch/lumberjack.v2 v2.2.1
go: downloading go.uber.org/atomic v1.11.0
go: downloading go.uber.org/multierr v1.11.0
go: downloading github.com/pingcap/failpoint v0.0.0-20220801062533-2eaa32854a6c
go: downloading github.com/pingcap/tidb/pkg/parser v0.0.0-20240410110152-5fc42c9be2f5
go: downloading github.com/coreos/go-semver v0.3.1
go: downloading github.com/go-sql-driver/mysql v1.7.1
go: downloading google.golang.org/grpc v1.62.1
/tmp/tidb_cdc_test/row_format/stdout.log
/tmp/tidb_cdc_test/row_format/tidb-slow.log
/tmp/tidb_cdc_test/row_format/tikv1.log
/tmp/tidb_cdc_test/row_format/tidb.log
/tmp/tidb_cdc_test/row_format/tiflash/db/proxy/db/000005.log
/tmp/tidb_cdc_test/row_format/tiflash/log/error.log
/tmp/tidb_cdc_test/row_format/tiflash/log/server.log
/tmp/tidb_cdc_test/row_format/tiflash/log/proxy.log
/tmp/tidb_cdc_test/row_format/tidb_down.log
/tmp/tidb_cdc_test/row_format/tikv1/db/000005.log
/tmp/tidb_cdc_test/row_format/pd1.log
/tmp/tidb_cdc_test/row_format/tidb_other.log
/tmp/tidb_cdc_test/row_format/down_pd/hot-region/000001.log
/tmp/tidb_cdc_test/row_format/down_pd/region-meta/000001.log
/tmp/tidb_cdc_test/row_format/tikv3.log
[2024/05/04 19:44:41.797 +08:00] [INFO] [main.go:178] ["73 insert success: 600"]
[2024/05/04 19:44:41.799 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:41.865 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:41.872 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:41.892 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:41.902 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:41.968 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:41.981 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:41.985 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:41.998 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
Logging trace to /tmp/tidb_cdc_test/changefeed_error/tiflash/log/server.log
Logging errors to /tmp/tidb_cdc_test/changefeed_error/tiflash/log/error.log
arg matches is ArgMatches { args: {"engine-version": MatchedArg { occurs: 1, indices: [12], vals: ["v8.2.0-alpha-16-g8e170090f"] }, "pd-endpoints": MatchedArg { occurs: 1, indices: [16], vals: ["127.0.0.1:2379"] }, "advertise-addr": MatchedArg { occurs: 1, indices: [4], vals: ["127.0.0.1:9000"] }, "addr": MatchedArg { occurs: 1, indices: [20], vals: ["127.0.0.1:9000"] }, "engine-addr": MatchedArg { occurs: 1, indices: [2], vals: ["127.0.0.1:9500"] }, "engine-git-hash": MatchedArg { occurs: 1, indices: [10], vals: ["8e170090fad91c94bef8d908e21c195c1d145b02"] }, "log-file": MatchedArg { occurs: 1, indices: [18], vals: ["/tmp/tidb_cdc_test/changefeed_error/tiflash/log/proxy.log"] }, "engine-label": MatchedArg { occurs: 1, indices: [14], vals: ["tiflash"] }, "config": MatchedArg { occurs: 1, indices: [8], vals: ["/tmp/tidb_cdc_test/changefeed_error/tiflash-proxy.toml"] }, "data-dir": MatchedArg { occurs: 1, indices: [6], vals: ["/tmp/tidb_cdc_test/changefeed_error/tiflash/db/proxy"] }}, subcommand: None, usage: Some("USAGE:\n    TiFlash Proxy [FLAGS] [OPTIONS] --engine-git-hash <engine-git-hash> --engine-label <engine-label> --engine-version <engine-version>") }
/tmp/tidb_cdc_test/row_format/cdc_kafka_consumer.log
/tmp/tidb_cdc_test/row_format/pd1/hot-region/000001.log
/tmp/tidb_cdc_test/row_format/pd1/region-meta/000001.log
/tmp/tidb_cdc_test/row_format/tikv2.log
/tmp/tidb_cdc_test/row_format/cdc_kafka_consumer_stdout.log
/tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0004/000002.log
/tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0003/000002.log
/tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0001/000002.log
/tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0005/000002.log
/tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0007/000002.log
/tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0006/000002.log
/tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0000/000002.log
/tmp/tidb_cdc_test/row_format/cdc_data/tmp/sorter/0002/000002.log
+ ls -alh log-G03.tar.gz
-rw-r--r--. 1 jenkins jenkins 4.8M May  4 19:44 log-G03.tar.gz
[2024/05/04 19:44:42.066 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:42.066 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:42.151 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:42.158 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:42.164 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:42.186 +08:00] [INFO] [main.go:178] ["72 insert success: 700"]
[2024/05/04 19:44:42.206 +08:00] [INFO] [main.go:178] ["73 insert success: 700"]
[2024/05/04 19:44:42.207 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:42.255 +08:00] [INFO] [main.go:88] ["testGetDefaultValue take 15.884980598s"]
[2024/05/04 19:44:42.276 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:42.277 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[Pipeline] // withEnv
[2024/05/04 19:44:42.301 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:42.302 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:42.369 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:42.396 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:42.400 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:42.402 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:42.472 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:42.482 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
go: downloading github.com/golang/protobuf v1.5.4
go: downloading golang.org/x/net v0.24.0
go: downloading golang.org/x/sys v0.19.0
go: downloading google.golang.org/protobuf v1.33.0
go: downloading google.golang.org/genproto/googleapis/rpc v0.0.0-20240401170217-c3f982113cda
go: downloading google.golang.org/genproto v0.0.0-20240401170217-c3f982113cda
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/ddl_only_block_related_table/run.sh using Sink-Type: kafka... <<=================
The 1 times to try to start tidb cluster...
[Pipeline] }
[Pipeline] archiveArtifacts
Archiving artifacts
go: downloading golang.org/x/text v0.14.0
[2024/05/04 19:44:42.564 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:42.566 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:42.577 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:42.593 +08:00] [INFO] [main.go:178] ["72 insert success: 800"]
[2024/05/04 19:44:42.650 +08:00] [INFO] [main.go:178] ["73 insert success: 800"]
[2024/05/04 19:44:42.653 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:42.687 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:42.689 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:42.706 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:42.746 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:42.775 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G01'
Sending interrupt signal to process
Killing processes
[2024/05/04 19:44:42.805 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:42.850 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:42.852 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
[2024/05/04 19:44:42.885 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:42.899 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:42.975 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:42.983 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:42.995 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:43.008 +08:00] [INFO] [main.go:178] ["72 insert success: 900"]
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.multi_capture.cli.9237.out cli tso query --pd=http://127.0.0.1:2379
Recording fingerprints
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
[2024/05/04 19:44:43.062 +08:00] [INFO] [main.go:178] ["73 insert success: 900"]
[2024/05/04 19:44:43.070 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:43.098 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:43.100 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:43.145 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:43.146 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:43.179 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:43.248 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:43.252 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:43.256 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:43.293 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:43.301 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
The 1 times to try to start tidb cluster...
[2024/05/04 19:44:43.446 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:43.448 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:43.466 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:43.480 +08:00] [INFO] [main.go:178] ["72 insert success: 1000"]
[2024/05/04 19:44:43.553 +08:00] [INFO] [main.go:178] ["73 insert success: 1000"]
[2024/05/04 19:44:43.561 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
script returned exit code 143
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0daec2ac0014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:15585, start at 2024-05-04 19:44:42.174461853 +0800 CST m=+5.128434108	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:42.180 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:42.155 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:42.155 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0daec2ac0014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:15585, start at 2024-05-04 19:44:42.174461853 +0800 CST m=+5.128434108	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:42.180 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:42.155 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:42.155 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Verifying Downstream TiDB is started...
VARIABLE_NAME	VARIABLE_VALUE	COMMENT
bootstrapped	True	Bootstrap flag. Do not delete.
tidb_server_version	196	Bootstrap version. Do not delete.
system_tz	Asia/Shanghai	TiDB Global System Timezone.
new_collation_enabled	True	If the new collations are enabled. Do not edit it.
ddl_table_version	3	DDL Table Version. Do not delete.
tikv_gc_leader_uuid	63d0daec1d00014	Current GC worker leader UUID. (DO NOT EDIT)
tikv_gc_leader_desc	host:pingcap-tiflow-pull-cdc-integration-kafka-test-1845-gc0bt-fzcb7, pid:15663, start at 2024-05-04 19:44:42.129765176 +0800 CST m=+5.034789964	Host name and pid of current GC leader. (DO NOT EDIT)
tikv_gc_leader_lease	20240504-19:46:42.136 +0800	Current GC worker leader lease. (DO NOT EDIT)
tikv_gc_auto_concurrency	true	Let TiDB pick the concurrency automatically. If set false, tikv_gc_concurrency will be used
tikv_gc_enable	true	Current GC enable status
tikv_gc_run_interval	10m0s	GC run interval, at least 10m, in Go format.
tikv_gc_life_time	10m0s	All versions within life time will not be collected by GC, at least 10m, in Go format.
tikv_gc_last_run_time	20240504-19:44:42.100 +0800	The time when last GC starts. (DO NOT EDIT)
tikv_gc_safe_point	20240504-19:34:42.100 +0800	All versions after safe point can be accessed. (DO NOT EDIT)
Starting Upstream TiFlash...
TiFlash
Release Version: v8.2.0-alpha-16-g8e170090f
Edition:         Community
Git Commit Hash: 8e170090fad91c94bef8d908e21c195c1d145b02
Git Branch:      HEAD
UTC Build Time:  2024-04-30 02:34:21
Enable Features: jemalloc sm4(GmSSL) avx2 avx512 unwind thinlto
Profile:         RELWITHDEBINFO
Compiler:        clang++ 13.0.0

Raft Proxy
Git Commit Hash:   7dc50b4eb06124e31f03adb06c20ff7ab61c5f79
Git Commit Branch: HEAD
UTC Build Time:    2024-04-30 02:38:45
Rust Version:      rustc 1.67.0-nightly (96ddd32c4 2022-11-14)
Storage Engine:    tiflash
Prometheus Prefix: tiflash_proxy_
Profile:           release
Enable Features:   external-jemalloc portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine cloud-aws cloud-gcp cloud-azure openssl-vendored
Verifying Upstream TiFlash is started...
script returned exit code 143
start tidb cluster in /tmp/tidb_cdc_test/changefeed_finish
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
start tidb cluster in /tmp/tidb_cdc_test/tiflash
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
[2024/05/04 19:44:43.590 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:43.594 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:43.648 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:43.650 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:43.677 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:43.706 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:43.747 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:43.750 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:43.777 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:43.781 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
script returned exit code 143
+ pd_host=127.0.0.1
+ pd_port=2379
+ is_tls=false
+ '[' false == true ']'
++ run_cdc_cli tso query --pd=http://127.0.0.1:2379
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.changefeed_error.cli.5708.out cli tso query --pd=http://127.0.0.1:2379
[2024/05/04 19:44:43.856 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:43.859 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:43.869 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:43.883 +08:00] [INFO] [main.go:178] ["72 insert success: 1100"]
[2024/05/04 19:44:43.944 +08:00] [INFO] [main.go:178] ["73 insert success: 1100"]
[2024/05/04 19:44:43.962 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:43.985 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:43.995 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:44.056 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:44.058 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
check diff failed 1-th time, retry later
[2024/05/04 19:44:44.086 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:44.161 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:44.170 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
[2024/05/04 19:44:44.171 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:44.197 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:44.202 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:44.275 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:44.278 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:44.286 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:44.303 +08:00] [INFO] [main.go:178] ["72 insert success: 1200"]
+ set +x
+ tso='449530582249242625
PASS
coverage: 1.8% of statements in github.com/pingcap/tiflow/...'
+ echo 449530582249242625 PASS coverage: 1.8% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
[2024/05/04 19:44:44.354 +08:00] [INFO] [main.go:178] ["73 insert success: 1200"]
[2024/05/04 19:44:44.369 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:44.394 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:44.403 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:44.465 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:44.465 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:44.493 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:44.571 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:44.577 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:44.580 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
go: downloading github.com/cznic/mathutil v0.0.0-20181122101859-297441e03548
go: downloading golang.org/x/exp v0.0.0-20240409090435-93d18d7e34b8
go: downloading github.com/ngaut/pools v0.0.0-20180318154953-b7bc8c42aac7
go: downloading github.com/jellydator/ttlcache/v3 v3.0.1
go: downloading github.com/pingcap/kvproto v0.0.0-20240227073058-929ab83f9754
go: downloading github.com/prometheus/client_golang v1.19.0
go: downloading github.com/tidwall/btree v1.7.0
go: downloading github.com/scalalang2/golang-fifo v0.1.5
go: downloading github.com/tikv/pd/client v0.0.0-20240322051414-fb9e2d561b6e
go: downloading github.com/tikv/client-go/v2 v2.0.8-0.20240409022718-714958ccd4d5
go: downloading golang.org/x/sync v0.7.0
go: downloading github.com/coocood/freecache v1.2.1
go: downloading github.com/pingcap/tipb v0.0.0-20240318032315-55a7867ddd50
go: downloading github.com/influxdata/tdigest v0.0.1
go: downloading github.com/google/btree v1.1.2
go: downloading go.etcd.io/etcd/client/v3 v3.5.12
go: downloading gopkg.in/yaml.v2 v2.4.0
go: downloading github.com/gorilla/mux v1.8.0
go: downloading github.com/pingcap/sysutil v1.0.1-0.20240311050922-ae81ee01f3a5
go: downloading github.com/docker/go-units v0.5.0
go: downloading github.com/grpc-ecosystem/go-grpc-middleware v1.4.0
go: downloading github.com/prometheus/client_model v0.6.1
go: downloading github.com/tiancaiamao/gp v0.0.0-20221230034425-4025bc8a4d4a
go: downloading github.com/google/uuid v1.6.0
go: downloading github.com/opentracing/opentracing-go v1.2.0
go: downloading github.com/danjacques/gofslock v0.0.0-20240212154529-d899e02bfe22
go: downloading github.com/shirou/gopsutil/v3 v3.24.2
go: downloading github.com/yangkeao/ldap/v3 v3.4.5-0.20230421065457-369a3bab1117
go: downloading github.com/gogo/protobuf v1.3.2
go: downloading github.com/opentracing/basictracer-go v1.1.0
go: downloading github.com/dolthub/swiss v0.2.1
go: downloading github.com/uber/jaeger-client-go v2.30.0+incompatible
go: downloading github.com/stretchr/testify v1.9.0
go: downloading github.com/twmb/murmur3 v1.1.6
go: downloading github.com/spf13/pflag v1.0.5
go: downloading go.etcd.io/etcd/api/v3 v3.5.12
go: downloading github.com/golang/snappy v0.0.4
go: downloading github.com/cockroachdb/errors v1.11.1
go: downloading golang.org/x/tools v0.20.0
go: downloading cloud.google.com/go/storage v1.39.1
go: downloading github.com/Azure/azure-sdk-for-go/sdk/azcore v1.9.1
go: downloading github.com/Azure/azure-sdk-for-go/sdk/azidentity v1.5.1
go: downloading github.com/Azure/azure-sdk-for-go/sdk/storage/azblob v1.0.0
go: downloading github.com/aliyun/alibaba-cloud-sdk-go v1.61.1581
go: downloading github.com/tikv/pd v1.1.0-beta.0.20240407022249-7179657d129b
go: downloading github.com/aws/aws-sdk-go v1.50.0
go: downloading github.com/go-resty/resty/v2 v2.11.0
go: downloading github.com/klauspost/compress v1.17.8
go: downloading github.com/ks3sdklib/aws-sdk-go v1.2.9
go: downloading cloud.google.com/go v0.112.2
go: downloading golang.org/x/oauth2 v0.18.0
go: downloading google.golang.org/api v0.170.0
go: downloading github.com/ngaut/sync2 v0.0.0-20141008032647-7a24ed77b2ef
go: downloading github.com/cespare/xxhash/v2 v2.3.0
go: downloading go.uber.org/mock v0.4.0
go: downloading github.com/cockroachdb/pebble v1.1.0
go: downloading github.com/jfcg/sorty/v2 v2.1.0
go: downloading github.com/carlmjohnson/flagext v0.21.0
go: downloading golang.org/x/time v0.5.0
go: downloading github.com/dgraph-io/ristretto v0.1.1
go: downloading github.com/Azure/go-ntlmssp v0.0.0-20221128193559-754e69321358
go: downloading github.com/go-asn1-ber/asn1-ber v1.5.4
go: downloading github.com/dolthub/maphash v0.1.0
go: downloading github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec
go: downloading go.etcd.io/etcd/client/pkg/v3 v3.5.12
go: downloading github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc
go: downloading github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2
go: downloading gopkg.in/yaml.v3 v3.0.1
go: downloading github.com/Azure/azure-sdk-for-go/sdk/internal v1.5.1
go: downloading github.com/AzureAD/microsoft-authentication-library-for-go v1.2.1
go: downloading golang.org/x/crypto v0.22.0
go: downloading github.com/beorn7/perks v1.0.1
go: downloading github.com/prometheus/common v0.52.2
go: downloading github.com/prometheus/procfs v0.13.0
go: downloading github.com/pkg/errors v0.9.1
go: downloading github.com/uber/jaeger-lib v2.4.1+incompatible
go: downloading github.com/cockroachdb/redact v1.1.5
go: downloading github.com/getsentry/sentry-go v0.27.0
go: downloading github.com/cockroachdb/logtags v0.0.0-20230118201751-21c54148d20b
go: downloading github.com/tklauser/go-sysconf v0.3.12
go: downloading github.com/joho/sqltocsv v0.0.0-20210428211105-a6d6801d59df
go: downloading github.com/jedib0t/go-pretty/v6 v6.2.2
go: downloading github.com/lestrrat-go/jwx/v2 v2.0.21
go: downloading github.com/dgryski/go-farm v0.0.0-20200201041132-a6ae2369ad13
go: downloading github.com/cloudfoundry/gosigar v1.3.6
go: downloading github.com/otiai10/copy v1.2.0
go: downloading github.com/asaskevich/govalidator v0.0.0-20230301143203-a9d515a09cc2
go: downloading github.com/spkg/bom v1.0.0
go: downloading github.com/xitongsys/parquet-go v1.6.0
go: downloading github.com/google/pprof v0.0.0-20240117000934-35fc243c5815
go: downloading github.com/wangjohn/quickselect v0.0.0-20161129230411-ed8402a42d5f
go: downloading google.golang.org/genproto/googleapis/api v0.0.0-20240401170217-c3f982113cda
go: downloading github.com/jfcg/sixb v1.3.8
go: downloading github.com/kr/pretty v0.3.1
go: downloading cloud.google.com/go/compute/metadata v0.2.3
go: downloading github.com/coreos/go-systemd/v22 v22.5.0
go: downloading github.com/cheggaaa/pb/v3 v3.0.8
go: downloading cloud.google.com/go/iam v1.1.7
go: downloading cloud.google.com/go/compute v1.25.1
go: downloading github.com/googleapis/gax-go/v2 v2.12.3
go: downloading github.com/robfig/cron/v3 v3.0.1
go: downloading github.com/pingcap/goleveldb v0.0.0-20191226122134-f82aafb29989
go: downloading github.com/pingcap/badger v1.5.1-0.20230103063557-828f39b09b6d
go: downloading github.com/robfig/cron v1.2.0
go: downloading github.com/tklauser/numcpus v0.6.1
go: downloading github.com/kylelemons/godebug v1.1.0
go: downloading github.com/pkg/browser v0.0.0-20240102092130-5ac0b6a4141c
go: downloading github.com/kr/text v0.2.0
go: downloading github.com/rogpeppe/go-internal v1.12.0
go: downloading github.com/VividCortex/ewma v1.2.0
go: downloading github.com/fatih/color v1.16.0
go: downloading github.com/mattn/go-colorable v0.1.13
go: downloading github.com/mattn/go-isatty v0.0.20
go: downloading github.com/mattn/go-runewidth v0.0.15
go: downloading go.opencensus.io v0.23.1-0.20220331163232-052120675fac
go: downloading go.opentelemetry.io/otel v1.24.0
go: downloading go.opentelemetry.io/otel/trace v1.24.0
go: downloading github.com/apache/thrift v0.16.0
go: downloading github.com/lestrrat-go/blackmagic v1.0.2
go: downloading github.com/lestrrat-go/httprc v1.0.5
go: downloading github.com/lestrrat-go/iter v1.0.2
go: downloading github.com/lestrrat-go/option v1.0.1
go: downloading github.com/dustin/go-humanize v1.0.1
go: downloading github.com/golang/glog v1.2.0
go: downloading github.com/golang-jwt/jwt/v5 v5.2.0
go: downloading github.com/rivo/uniseg v0.4.7
go: downloading github.com/lestrrat-go/httpcc v1.0.1
***************** properties *****************
"insertproportion"="0"
"mysql.port"="4000"
"dotransactions"="false"
"mysql.host"="127.0.0.1"
"scanproportion"="0"
"threadcount"="2"
"operationcount"="0"
"updateproportion"="0"
"mysql.db"="multi_capture_1"
"readallfields"="true"
"readproportion"="0"
"workload"="core"
"requestdistribution"="uniform"
"recordcount"="10"
"mysql.user"="root"
**********************************************
Run finished, takes 9.135197ms
INSERT - Takes(s): 0.0, Count: 10, OPS: 1923.1, Avg(us): 1753, Min(us): 1176, Max(us): 3920, 95th(us): 4000, 99th(us): 4000
[2024/05/04 19:44:44.605 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:44.645 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:44.684 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:44.690 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:44.705 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:44.752 +08:00] [INFO] [main.go:178] ["72 insert success: 1300"]
[2024/05/04 19:44:44.759 +08:00] [INFO] [main.go:178] ["73 insert success: 1300"]
[2024/05/04 19:44:44.776 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:44.807 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:44.812 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
start tidb cluster in /tmp/tidb_cdc_test/new_ci_collation
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
go: downloading github.com/ncw/directio v1.0.5
go: downloading github.com/coocood/rtutil v0.0.0-20190304133409-c84515f646f2
go: downloading github.com/coocood/bbloom v0.0.0-20190830030839-58deb6228d64
go: downloading github.com/klauspost/cpuid v1.3.1
go: downloading github.com/golang-jwt/jwt v3.2.2+incompatible
go: downloading github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da
go: downloading github.com/go-logr/logr v1.4.1
go: downloading go.opentelemetry.io/otel/metric v1.24.0
go: downloading github.com/go-logr/stdr v1.2.2
***************** properties *****************
"threadcount"="2"
"mysql.port"="4000"
"recordcount"="10"
"insertproportion"="0"
"workload"="core"
"readproportion"="0"
"mysql.user"="root"
"readallfields"="true"
"scanproportion"="0"
"mysql.host"="127.0.0.1"
"requestdistribution"="uniform"
"mysql.db"="multi_capture_2"
"dotransactions"="false"
"operationcount"="0"
"updateproportion"="0"
**********************************************
Run finished, takes 9.202801ms
INSERT - Takes(s): 0.0, Count: 10, OPS: 1946.7, Avg(us): 1768, Min(us): 1052, Max(us): 3981, 95th(us): 4000, 99th(us): 4000
[2024/05/04 19:44:44.868 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:44.869 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:44.896 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:44.974 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:44.978 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:44.983 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:45.052 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:45.056 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:45.088 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
go: downloading github.com/DataDog/zstd v1.5.5
go: downloading github.com/cockroachdb/tokenbucket v0.0.0-20230807174530-cc333fc44b06
script returned exit code 143
***************** properties *****************
"requestdistribution"="uniform"
"operationcount"="0"
"mysql.db"="multi_capture_3"
"dotransactions"="false"
"updateproportion"="0"
"scanproportion"="0"
"mysql.port"="4000"
"readallfields"="true"
"threadcount"="2"
"workload"="core"
"mysql.host"="127.0.0.1"
"mysql.user"="root"
"recordcount"="10"
"insertproportion"="0"
"readproportion"="0"
**********************************************
Run finished, takes 10.056949ms
INSERT - Takes(s): 0.0, Count: 10, OPS: 1724.0, Avg(us): 1917, Min(us): 1232, Max(us): 4114, 95th(us): 5000, 99th(us): 5000
***************** properties *****************
"recordcount"="10"
"mysql.port"="4000"
"insertproportion"="0"
"mysql.user"="root"
"threadcount"="2"
"dotransactions"="false"
"mysql.host"="127.0.0.1"
"operationcount"="0"
"readallfields"="true"
"scanproportion"="0"
"updateproportion"="0"
"workload"="core"
"readproportion"="0"
"mysql.db"="multi_capture_4"
"requestdistribution"="uniform"
**********************************************
[2024/05/04 19:44:45.099 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:45.151 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:45.169 +08:00] [INFO] [main.go:178] ["72 insert success: 1400"]
[2024/05/04 19:44:45.170 +08:00] [INFO] [main.go:178] ["73 insert success: 1400"]
[2024/05/04 19:44:45.188 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:45.263 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:45.269 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:45.296 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:45.345 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
Run finished, takes 9.771074ms
INSERT - Takes(s): 0.0, Count: 10, OPS: 1638.0, Avg(us): 1823, Min(us): 1138, Max(us): 3632, 95th(us): 4000, 99th(us): 4000
[Sat May  4 19:44:45 CST 2024] <<<<<< START cdc server in multi_capture case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8301/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ (( i = 0 ))
+ (( i <= 50 ))
+ GO_FAILPOINTS=
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.multi_capture.93769378.out server --log-file /tmp/tidb_cdc_test/multi_capture/cdc1.log --log-level debug --data-dir /tmp/tidb_cdc_test/multi_capture/cdc_data1 --cluster-id default --addr 127.0.0.1:8301
++ curl -vsL --max-time 20 http://127.0.0.1:8301/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8301 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8301; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
Verifying downstream PD is started...
[2024/05/04 19:44:45.372 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:45.461 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:45.468 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:45.472 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:45.564 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:45.565 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:45.760 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
check diff failed 2-th time, retry later
[2024/05/04 19:44:45.855 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:45.861 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:45.950 +08:00] [INFO] [main.go:178] ["72 insert success: 1500"]
[2024/05/04 19:44:45.956 +08:00] [INFO] [main.go:178] ["73 insert success: 1500"]
[2024/05/04 19:44:45.976 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
Verifying downstream PD is started...
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/processor_etcd_worker_delay/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:44:45 CST 2024] <<<<<< run test case processor_etcd_worker_delay success! >>>>>>
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
[2024/05/04 19:44:46.167 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:46.179 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:46.269 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:46.345 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:46.359 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:46.364 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:46.372 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:46.386 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:46.456 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
[2024/05/04 19:44:46.472 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:46.473 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:46.504 +08:00] [INFO] [main.go:178] ["72 insert success: 1600"]
[2024/05/04 19:44:46.509 +08:00] [INFO] [main.go:178] ["73 insert success: 1600"]
[2024/05/04 19:44:46.515 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:46.602 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:46.606 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
start tidb cluster in /tmp/tidb_cdc_test/many_pk_or_uk
Starting Upstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Starting Downstream PD...
Release Version: v8.2.0-alpha-14-g1679dbca2
Edition: Community
Git Commit Hash: 1679dbca25b3483d1375c7e747da27e99ad77360
Git Branch: master
UTC Build Time:  2024-04-30 08:09:12
Verifying upstream PD is started...
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
[2024/05/04 19:44:46.660 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:46.711 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:46.713 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:46.754 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:46.763 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
[2024/05/04 19:44:46.769 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:46.804 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[2024/05/04 19:44:46.864 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:46.867 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
\033[0;36m<<< Run all test success >>>\033[0m
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
[2024/05/04 19:44:46.876 +08:00] [INFO] [main.go:178] ["73 insert success: 1700"]
[2024/05/04 19:44:46.883 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
[2024/05/04 19:44:46.897 +08:00] [INFO] [main.go:178] ["72 insert success: 1700"]
[2024/05/04 19:44:46.977 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[2024/05/04 19:44:46.986 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
[2024/05/04 19:44:47.005 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
[2024/05/04 19:44:47.093 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
{"level":"warn","ts":1714823087.129507,"caller":"v3@v3.5.12/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00202c700/127.0.0.1:2379","attempt":0,"error":"rpc error: code = Unavailable desc = error reading from server: EOF"}
script returned exit code 143
[2024/05/04 19:44:47.162 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:47.175 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[2024/05/04 19:44:47.179 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
[2024/05/04 19:44:47.198 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
[2024/05/04 19:44:47.254 +08:00] [INFO] [main.go:178] ["73 insert success: 2000"]
[2024/05/04 19:44:47.296 +08:00] [INFO] [main.go:178] ["73 insert success: 1800"]
[2024/05/04 19:44:47.300 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
[2024/05/04 19:44:47.300 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[2024/05/04 19:44:47.331 +08:00] [INFO] [main.go:178] ["72 insert success: 1800"]
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
[2024/05/04 19:44:47.382 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[2024/05/04 19:44:47.472 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[2024/05/04 19:44:47.551 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
[2024/05/04 19:44:47.569 +08:00] [INFO] [main.go:178] ["73 insert success: 2000"]
[2024/05/04 19:44:47.576 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
[2024/05/04 19:44:47.596 +08:00] [INFO] [main.go:178] ["72 insert success: 2000"]
script returned exit code 143
[2024/05/04 19:44:47.746 +08:00] [INFO] [main.go:178] ["73 insert success: 1900"]
[2024/05/04 19:44:47.756 +08:00] [INFO] [main.go:178] ["73 insert success: 2000"]
[2024/05/04 19:44:47.764 +08:00] [INFO] [main.go:178] ["72 insert success: 2000"]
[2024/05/04 19:44:47.768 +08:00] [INFO] [main.go:178] ["72 insert success: 1900"]
[2024/05/04 19:44:47.806 +08:00] [INFO] [main.go:178] ["73 insert success: 2000"]
[2024/05/04 19:44:47.867 +08:00] [INFO] [main.go:178] ["73 insert success: 2000"]
check diff failed 3-th time, retry later
script returned exit code 143
Starting Upstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.2.0-alpha
Edition:           Community
Git Commit Hash:   72a0fd5b00235a7c56014b77ddd933e2a0d33c88
Git Commit Branch: master
UTC Build Time:    2024-04-30 02:23:51
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   memory-engine pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
[2024/05/04 19:44:47.897 +08:00] [INFO] [main.go:178] ["72 insert success: 2000"]
[2024/05/04 19:44:47.922 +08:00] [INFO] [main.go:178] ["72 insert success: 2000"]
[2024/05/04 19:44:47.949 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLsb12a3c01_becc_4017_ae72_d0e93de15beb"]
[2024/05/04 19:44:48.044 +08:00] [INFO] [main.go:178] ["73 insert success: 2000"]
[2024/05/04 19:44:48.066 +08:00] [INFO] [main.go:178] ["72 insert success: 2000"]
[2024/05/04 19:44:48.070 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs0f3c49fb_1d3b_443f_8f46_85ef2027ae47"]
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
[2024/05/04 19:44:48.180 +08:00] [INFO] [main.go:178] ["73 insert success: 2100"]
[2024/05/04 19:44:48.248 +08:00] [INFO] [main.go:178] ["72 insert success: 2100"]
[2024/05/04 19:44:48.274 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
[2024/05/04 19:44:48.278 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:48.347 +08:00] [INFO] [main.go:178] ["73 insert success: 2100"]
[2024/05/04 19:44:48.349 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs6b780135_2ddf_4f75_b127_4c076a8dddbd"]
[2024/05/04 19:44:48.370 +08:00] [INFO] [main.go:178] ["72 insert success: 2100"]
[2024/05/04 19:44:48.371 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLs16aab79b_2873_4bc4_bb5e_2aff57fe10fd"]
[2024/05/04 19:44:48.378 +08:00] [INFO] [main.go:178] ["72 insert success: 100"]
[2024/05/04 19:44:48.380 +08:00] [INFO] [main.go:178] ["73 insert success: 100"]
{"level":"warn","ts":1714823088.2572474,"caller":"v3@v3.5.12/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc002494380/127.0.0.1:2379","attempt":0,"error":"rpc error: code = Unavailable desc = error reading from server: EOF"}
script returned exit code 143
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
[2024/05/04 19:44:48.496 +08:00] [INFO] [main.go:835] ["running ddl test: testMultiDDLsfc2f058f_ac31_43ac_95a9_ecc2a8fb5e17"]
{"level":"warn","ts":1714823088.5464678,"caller":"v3@v3.5.12/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00217e540/127.0.0.1:2379","attempt":0,"error":"rpc error: code = Unavailable desc = error reading from server: read tcp 127.0.0.1:45040->127.0.0.1:2379: read: connection reset by peer"}
script returned exit code 143
Starting Upstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
Starting Upstream TiDB...
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
script returned exit code 143
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.2.0-alpha-79-g600b2ed4bf
Edition: Community
Git Commit Hash: 600b2ed4bf0aa38224a1c4c4c68831820735515c
Git Branch: master
UTC Build Time: 2024-05-01 02:56:48
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
script returned exit code 143
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
script returned exit code 143
script returned exit code 143
kill finished with exit code 0
Killing processes
find: '/tmp/tidb_cdc_test/*/*': No such file or directory
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/tests/integration_tests/sink_hang/run.sh using Sink-Type: kafka... <<=================
[Sat May  4 19:44:50 CST 2024] <<<<<< run test case sink_hang success! >>>>>>
script returned exit code 143
kill finished with exit code 1
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (ws/jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1845/tiflow-cdc already exists)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
script returned exit code 143
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // cache
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // dir
[Pipeline] }
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G03'
[Pipeline] // node
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G00'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G02'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G04'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G05'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G06'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G07'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G08'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G09'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G10'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G12'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G13'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G14'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G15'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G16'
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G11'
[Pipeline] // parallel
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] End of Pipeline
ERROR: script returned exit code 1
Finished: FAILURE