Skip to content

Console Output

Started by user Jenkins Admin
Obtained pipelines/pingcap/tiflow/latest/pull_cdc_integration_kafka_test.groovy from git https://github.com/PingCAP-QE/ci.git
Loading library tipipeline@main
Library tipipeline@main is cached. Copying from home.
[Pipeline] Start of Pipeline
[Pipeline] readJSON
[Pipeline] readTrusted
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_kafka_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-kafka-test-1846-fvtv0-85pjh
Agent pingcap-tiflow-pull-cdc-integration-kafka-test-1846-fvtv0-85pjh is provisioned from template pingcap_tiflow_pull_cdc_integration_kafka_test_1846-fvtv0-glmv9
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1846/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1846/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "bd127dd805e6e3b8a98bfe542c5fa9f0f4e19f11"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_kafka_test_1846-fvtv0"
  name: "pingcap-tiflow-pull-cdc-integration-kafka-test-1846-fvtv0-85pjh"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "wurstmeister/zookeeper"
    imagePullPolicy: "IfNotPresent"
    name: "zookeeper"
    resources:
      limits:
        cpu: "2000m"
        memory: "4Gi"
      requests:
        cpu: "2000m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - args:
    - "cat"
    image: "hub.pingcap.net/jenkins/golang-tini:1.21"
    imagePullPolicy: "Always"
    name: "golang"
    resources:
      limits:
        cpu: "12"
        memory: "32Gi"
      requests:
        cpu: "12"
        memory: "32Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_CREATE_TOPICS"
      value: "big-message-test:1:1"
    - name: "KAFKA_BROKER_ID"
      value: "1"
    - name: "KAFKA_SSL_KEYSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_ZOOKEEPER_CONNECT"
      value: "localhost:2181"
    - name: "KAFKA_MESSAGE_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_REPLICA_FETCH_MAX_BYTES"
      value: "11534336"
    - name: "KAFKA_ADVERTISED_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "ZK"
      value: "zk"
    - name: "KAFKA_SSL_KEYSTORE_LOCATION"
      value: "/tmp/kafka.server.keystore.jks"
    - name: "KAFKA_SSL_KEY_PASSWORD"
      value: "test1234"
    - name: "KAFKA_SSL_TRUSTSTORE_PASSWORD"
      value: "test1234"
    - name: "KAFKA_LISTENERS"
      value: "SSL://127.0.0.1:9093,PLAINTEXT://127.0.0.1:9092"
    - name: "KAFKA_SSL_TRUSTSTORE_LOCATION"
      value: "/tmp/kafka.server.truststore.jks"
    - name: "RACK_COMMAND"
      value: "curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.keystore.jks\
        \ -o /tmp/kafka.server.keystore.jks && curl -sfL https://github.com/pingcap/tiflow/raw/6e62afcfecc4e3965d8818784327d4bf2600d9fa/tests/_certificates/kafka.server.truststore.jks\
        \ -o /tmp/kafka.server.truststore.jks"
    image: "wurstmeister/kafka:2.12-2.4.1"
    imagePullPolicy: "IfNotPresent"
    name: "kafka"
    resources:
      limits:
        cpu: "4000m"
        memory: "6Gi"
      requests:
        cpu: "4000m"
        memory: "6Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "KAFKA_SERVER"
      value: "127.0.0.1:9092"
    - name: "ZOOKEEPER_SERVER"
      value: "127.0.0.1:2181"
    - name: "DOWNSTREAM_DB_HOST"
      value: "127.0.0.1"
    - name: "USE_FLAT_MESSAGE"
      value: "true"
    - name: "DOWNSTREAM_DB_PORT"
      value: "3306"
    - name: "DB_NAME"
      value: "test"
    image: "rustinliu/ticdc-canal-json-adapter:latest"
    imagePullPolicy: "IfNotPresent"
    name: "canal-adapter"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/tmp"
      name: "volume-0"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "MYSQL_ROOT_PASSWORD"
      value: ""
    - name: "MYSQL_USER"
      value: "mysqluser"
    - name: "MYSQL_PASSWORD"
      value: "mysqlpw"
    - name: "MYSQL_ALLOW_EMPTY_PASSWORD"
      value: "yes"
    - name: "MYSQL_TCP_PORT"
      value: "3310"
    image: "quay.io/debezium/example-mysql:2.4"
    imagePullPolicy: "IfNotPresent"
    name: "mysql"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "BOOTSTRAP_SERVERS"
      value: "127.0.0.1:9092"
    - name: "GROUP_ID"
      value: "1"
    - name: "CONFIG_STORAGE_TOPIC"
      value: "my_connect_configs"
    - name: "OFFSET_STORAGE_TOPIC"
      value: "my_connect_offsets"
    - name: "STATUS_STORAGE_TOPIC"
      value: "my_connect_statuses"
    image: "quay.io/debezium/connect:2.4"
    name: "connect"
    resources:
      requests:
        cpu: "200m"
        memory: "4Gi"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1846-fvtv0-85pjh"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-kafka-test-1846-fvtv0-85pjh"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "volume-0"
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-kafka-test-1846-fvtv0-85pjh in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
[Pipeline] {
[Pipeline] stage
[Pipeline] { (Declarative: Checkout SCM)
[Pipeline] checkout
The recommended git tool is: git
No credentials specified
Cloning the remote Git repository
Using shallow clone with depth 1
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
 > git rev-list --no-walk 03312178c534dce949face80c69812d989e55009 # timeout=10
[Pipeline] }
[Pipeline] // stage
[Pipeline] withEnv
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 1 hr 5 min
[Pipeline] {
[Pipeline] stage
[Pipeline] { (Debug info)
[Pipeline] sh
+ printenv
PROW_JOB_ID=635325f1-8304-4ada-a204-8d3fd0ec56f8
JENKINS_NODE_COOKIE=e71101f3-6a77-46c5-9871-40130d996892
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1846/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1846-fvtv0-85pjh
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT=tcp://10.233.0.1:443
KUBERNETES_PORT_443_TCP_PORT=443
TERM=xterm
STAGE_NAME=Debug info
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_kafka_test-1846
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_kafka_test","buildid":"1786740771949383683","prowjobid":"635325f1-8304-4ada-a204-8d3fd0ec56f8","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"ccb0e990ebd1b8f66e79e01adf7fce4af132ff7b","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/ccb0e990ebd1b8f66e79e01adf7fce4af132ff7b","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1846/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1846/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1846/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
_=/usr/bin/printenv
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test
HUDSON_URL=https://do.pingcap.net/jenkins/
JOB_NAME=pingcap/tiflow/pull_cdc_integration_kafka_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1846
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786740771949383683
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=pull_cdc_integration_kafka_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/1846/display/redirect?page=tests
SHLVL=3
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_kafka_test_1846-fvtv0
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
TINI_VERSION=v0.19.0
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-236df335481f9578f70eb859f68d5ceead3aa27f6c9385fda1ec4c08661c0305
NODE_LABELS=pingcap_tiflow_pull_cdc_integration_kafka_test_1846-fvtv0 pingcap-tiflow-pull-cdc-integration-kafka-test-1846-fvtv0-85pjh
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-kafka-test-1846-fvtv0-85pjh
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_kafka_test/display/redirect
BUILD_NUMBER=1846
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
+ echo -------------------------
-------------------------
+ go env
GO111MODULE=''
GOARCH='amd64'
GOBIN=''
GOCACHE='/home/jenkins/.cache/go-build'
GOENV='/home/jenkins/.config/go/env'
GOEXE=''
GOEXPERIMENT=''
GOFLAGS=''
GOHOSTARCH='amd64'
GOHOSTOS='linux'
GOINSECURE=''
GOMODCACHE='/go/pkg/mod'
GONOPROXY=''
GONOSUMDB=''
GOOS='linux'
GOPATH='/go'
GOPRIVATE=''
GOPROXY='http://goproxy.apps.svc,https://proxy.golang.org,direct'
GOROOT='/usr/local/go'
GOSUMDB='sum.golang.org'
GOTMPDIR=''
GOTOOLCHAIN='auto'
GOTOOLDIR='/usr/local/go/pkg/tool/linux_amd64'
GOVCS=''
GOVERSION='go1.21.0'
GCCGO='gccgo'
GOAMD64='v1'
AR='ar'
CC='gcc'
CXX='g++'
CGO_ENABLED='1'
GOMOD='/dev/null'
GOWORK=''
CGO_CFLAGS='-O2 -g'
CGO_CPPFLAGS=''
CGO_CXXFLAGS='-O2 -g'
CGO_FFLAGS='-O2 -g'
CGO_LDFLAGS='-O2 -g'
PKG_CONFIG='pkg-config'
GOGCCFLAGS='-fPIC -m64 -pthread -Wl,--no-gc-sections -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build215593242=/tmp/go-build -gno-record-gcc-switches'
+ echo -------------------------
-------------------------
+ echo 'debug command: kubectl -n jenkins-tiflow exec -ti pingcap-tiflow-pull-cdc-integration-kafka-test-1846-fvtv0-85pjh bash'
debug command: kubectl -n jenkins-tiflow exec -ti pingcap-tiflow-pull-cdc-integration-kafka-test-1846-fvtv0-85pjh bash
[Pipeline] container
[Pipeline] {
[Pipeline] sh
+ dig github.com

; <<>> DiG 9.18.16 <<>> github.com
;; global options: +cmd
;; Got answer:
;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 20291
;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 1

;; OPT PSEUDOSECTION:
; EDNS: version: 0, flags:; udp: 1232
; COOKIE: 6fbfc1045bdf4512 (echoed)
;; QUESTION SECTION:
;github.com.			IN	A

;; ANSWER SECTION:
github.com.		18	IN	A	20.205.243.166

;; Query time: 0 msec
;; SERVER: 169.254.25.10#53(169.254.25.10) (UDP)
;; WHEN: Sat May 04 12:53:24 UTC 2024
;; MSG SIZE  rcvd: 77

[Pipeline] script
[Pipeline] {
[Pipeline] }
[Pipeline] // script
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Check diff files)
[Pipeline] container
[Pipeline] {
[Pipeline] script
[Pipeline] {
[Pipeline] withCredentials
Masking supported pattern matches of $token
[Pipeline] {
[Pipeline] httpRequest
Warning: A secret was passed to "httpRequest" using Groovy String interpolation, which is insecure.
		 Affected argument(s) used the following variable(s): [token]
		 See https://jenkins.io/redirect/groovy-string-interpolation for details.
HttpMethod: GET
URL: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=1&per_page=100
Content-Type: application/json
Authorization: *****
Sending request to url: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=1&per_page=100
Response Code: HTTP/1.1 200 OK
Success: Status code 200 is in the accepted range: 100:399
[Pipeline] httpRequest
Warning: A secret was passed to "httpRequest" using Groovy String interpolation, which is insecure.
		 Affected argument(s) used the following variable(s): [token]
		 See https://jenkins.io/redirect/groovy-string-interpolation for details.
HttpMethod: GET
URL: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=2&per_page=100
Content-Type: application/json
Authorization: *****
Sending request to url: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=2&per_page=100
Response Code: HTTP/1.1 200 OK
Success: Status code 200 is in the accepted range: 100:399
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] echo
pr_diff_files: [cdc/model/kv.go, cdc/model/sink.go, cdc/model/sink_test.go, cdc/processor/processor.go, cdc/processor/sinkmanager/manager.go, cdc/processor/sourcemanager/manager.go, cdc/redo/reader/reader.go, cdc/sink/dmlsink/factory/factory.go, cdc/sink/dmlsink/txn/mysql/dml.go, cdc/sink/dmlsink/txn/mysql/mysql.go, cdc/sink/dmlsink/txn/mysql/mysql_test.go, cmd/kafka-consumer/main.go, cmd/pulsar-consumer/main.go, cmd/storage-consumer/main.go, errors.toml, pkg/applier/redo.go, pkg/applier/redo_test.go, pkg/errors/cdc_errors.go, pkg/errors/helper.go, pkg/sink/codec/open/open_protocol_decoder.go, pkg/sink/codec/open/open_protocol_message.go, tests/integration_tests/_utils/check_sync_diff, tests/integration_tests/changefeed_dup_error_restart/conf/diff_config.toml, tests/integration_tests/changefeed_dup_error_restart/conf/workload, tests/integration_tests/changefeed_dup_error_restart/run.sh, tests/integration_tests/force_replicate_table/run.sh, tests/integration_tests/open_protocol_handle_key_only/data/data.sql, tests/integration_tests/open_protocol_handle_key_only/run.sh, tests/integration_tests/run_group.sh]
[Pipeline] echo
diff file not matched: cdc/model/kv.go
[Pipeline] }
[Pipeline] // script
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Checkout)
[Pipeline] timeout
Timeout set to expire in 10 min
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow
[Pipeline] {
[Pipeline] cache
Cache restored successfully (git/pingcap/tiflow/rev-be15534)
203635712 bytes in 1.29 secs (157761670 bytes/sec)
[Pipeline] {
[Pipeline] retry
[Pipeline] {
[Pipeline] script
[Pipeline] {
[Pipeline] sh
git version 2.36.6
Reinitialized existing Git repository in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/tiflow/.git/
.git
HEAD is now at be1553484 codec(ticdc): avro simplify the unit test (#11010)
POST git-upload-pack (656 bytes)
POST git-upload-pack (973 bytes)
From https://github.com/pingcap/tiflow
 = [up to date]          master               -> origin/master
 * [new ref]             refs/pull/10919/head -> origin/pr/10919/head
HEAD is now at be1553484 codec(ticdc): avro simplify the unit test (#11010)
๐Ÿšง Checkouting to base SHA:be1553484fe4c03594eabb8d7435c694e5fd7224...
HEAD is now at be1553484 codec(ticdc): avro simplify the unit test (#11010)
โœ… Checked. ๐ŸŽ‰
๐Ÿงพ HEAD info:
be1553484fe4c03594eabb8d7435c694e5fd7224
be1553484 codec(ticdc): avro simplify the unit test (#11010)
2a7a65c6f Support Sequences (#10203)
36e9e1bf6 cli(ticdc): allow client authentication to be enabled without tls (#11005)
๐Ÿšง Pre-merge heads of pull requests to base SHA: be1553484fe4c03594eabb8d7435c694e5fd7224 ...
Updating be1553484..ccb0e990e
Fast-forward
 cdc/model/kv.go                                    |   5 +
 cdc/model/sink.go                                  |  38 ++-
 cdc/model/sink_test.go                             |   9 +-
 cdc/processor/processor.go                         |  21 +-
 cdc/processor/sinkmanager/manager.go               |   5 +
 cdc/processor/sourcemanager/manager.go             |  66 +++-
 cdc/redo/reader/reader.go                          |  21 +-
 cdc/sink/dmlsink/factory/factory.go                |   8 +-
 cdc/sink/dmlsink/txn/mysql/dml.go                  |  23 ++
 cdc/sink/dmlsink/txn/mysql/mysql.go                |  89 +++---
 cdc/sink/dmlsink/txn/mysql/mysql_test.go           |   2 +-
 cmd/kafka-consumer/main.go                         |   4 +-
 cmd/pulsar-consumer/main.go                        |  17 +-
 cmd/storage-consumer/main.go                       |   4 +-
 errors.toml                                        |   5 +
 pkg/applier/redo.go                                | 303 +++++++++++++++++-
 pkg/applier/redo_test.go                           | 347 ++++++++++++++++++++-
 pkg/errors/cdc_errors.go                           |   4 +
 pkg/errors/helper.go                               |  19 ++
 pkg/sink/codec/open/open_protocol_decoder.go       |   1 +
 pkg/sink/codec/open/open_protocol_message.go       |   3 +
 tests/integration_tests/_utils/check_sync_diff     |   2 +-
 .../conf/diff_config.toml                          |  29 ++
 .../changefeed_dup_error_restart/conf/workload     |  13 +
 .../changefeed_dup_error_restart/run.sh            |  54 ++++
 .../integration_tests/force_replicate_table/run.sh |   4 +-
 .../open_protocol_handle_key_only/data/data.sql    |   2 +-
 .../open_protocol_handle_key_only/run.sh           |   4 +-
 tests/integration_tests/run_group.sh               |   5 +-
 29 files changed, 1002 insertions(+), 105 deletions(-)
 create mode 100644 tests/integration_tests/changefeed_dup_error_restart/conf/diff_config.toml
 create mode 100644 tests/integration_tests/changefeed_dup_error_restart/conf/workload
 create mode 100755 tests/integration_tests/changefeed_dup_error_restart/run.sh
๐Ÿงพ Pre-merged result:
ccb0e990ebd1b8f66e79e01adf7fce4af132ff7b
ccb0e990e f
57a3c8e82 f
b2edc574f f
โœ… Pre merged ๐ŸŽ‰
โœ… ~~~~~All done.~~~~~~
[Pipeline] }
[Pipeline] // script
[Pipeline] }
[Pipeline] // retry
[Pipeline] }
Cache saved successfully (git/pingcap/tiflow/rev-be15534-ccb0e99)
203866112 bytes in 8.18 secs (24929010 bytes/sec)
[Pipeline] // cache
[Pipeline] }
[Pipeline] // dir
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (prepare)
[Pipeline] timeout
Timeout set to expire in 20 min
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_kafka_test/third_party_download
[Pipeline] {
[Pipeline] retry
[Pipeline] {
[Pipeline] sh
+ cd ../tiflow
+ ./scripts/download-integration-test-binaries.sh master
Download binaries...
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0   1270      0 --:--:-- --:--:-- --:--:--  1281
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0    737      0 --:--:-- --:--:-- --:--:--   745
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0    755      0 --:--:-- --:--:-- --:--:--   759
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0   1082      0 --:--:-- --:--:-- --:--:--  1108
>>>
download tidb-server.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/tidb/600b2ed4bf0aa38224a1c4c4c68831820735515c/centos7/tidb-server.tar.gz
Aborted by Jenkins Admin
Sending interrupt signal to process
Killing processes
kill finished with exit code 0
script returned exit code 143
[Pipeline] }
[Pipeline] // retry
[Pipeline] }
[Pipeline] // dir
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Tests)
Stage "Tests" skipped due to earlier failure(s)
[Pipeline] parallel
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G00')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G01')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G02')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G03')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G04')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G05')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G06')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G07')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G08')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G09')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G10')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G11')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G12')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G13')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G14')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G15')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G16')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G17')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G00')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G01')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G02')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G03')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G04')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G05')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G06')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G07')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G08')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G09')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G10')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G11')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G12')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G13')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G14')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G15')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G16')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G17')
Stage "Matrix - TEST_GROUP = 'G00'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G01'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G02'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G03'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G04'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G05'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G06'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G07'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G08'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G09'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G10'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G11'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G12'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G13'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G14'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G15'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G16'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G17'" skipped due to earlier failure(s)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
Stage "Matrix - TEST_GROUP = 'G00'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G01'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G02'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G03'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G04'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G05'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G06'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G07'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G08'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G09'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G10'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G11'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G12'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G13'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G14'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G15'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G16'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G17'" skipped due to earlier failure(s)
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G00'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G01'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G02'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G03'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G04'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G05'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G06'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G07'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G08'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G09'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G10'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G11'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G12'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G13'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G14'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G15'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G16'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G17'
[Pipeline] // parallel
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] End of Pipeline
org.jenkinsci.plugins.workflow.actions.ErrorAction$ErrorId: e10525e3-beb4-4b7f-9015-7c2990733d04
Finished: ABORTED