Skip to content

Console Output

Started by user Jenkins Admin
Obtained pipelines/pingcap/tiflow/latest/pull_cdc_integration_storage_test.groovy from git https://github.com/PingCAP-QE/ci.git
Loading library tipipeline@main
Library tipipeline@main is cached. Copying from home.
[Pipeline] Start of Pipeline
[Pipeline] readJSON
[Pipeline] readTrusted
Obtained pipelines/pingcap/tiflow/latest/pod-pull_cdc_integration_storage_test.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Created Pod: kubernetes jenkins-tiflow/pingcap-tiflow-pull-cdc-integration-storage-test-1763-42t-t5xkh
Agent pingcap-tiflow-pull-cdc-integration-storage-test-1763-42t-t5xkh is provisioned from template pingcap_tiflow_pull_cdc_integration_storage_test_1763-42tt5-hfnpx
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_storage_test/1763/"
    runUrl: "job/pingcap/job/tiflow/job/pull_cdc_integration_storage_test/1763/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "02c356f35db46c574659cb8a50e55ce721dd691b"
    jenkins/label: "pingcap_tiflow_pull_cdc_integration_storage_test_1763-42tt5"
  name: "pingcap-tiflow-pull-cdc-integration-storage-test-1763-42t-t5xkh"
  namespace: "jenkins-tiflow"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
  containers:
  - image: "hub.pingcap.net/jenkins/centos7_golang-1.21:latest"
    name: "golang"
    resources:
      limits:
        memory: "16Gi"
        cpu: "6"
      requests:
        memory: "12Gi"
        cpu: "4"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      limits:
        memory: "256Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-storage-test-1763-42t-t5xkh"
    - name: "JENKINS_NAME"
      value: "pingcap-tiflow-pull-cdc-integration-storage-test-1763-42t-t5xkh"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir:
      medium: ""
    name: "workspace-volume"

Running on pingcap-tiflow-pull-cdc-integration-storage-test-1763-42t-t5xkh in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_storage_test
[Pipeline] {
[Pipeline] stage
[Pipeline] { (Declarative: Checkout SCM)
[Pipeline] checkout
The recommended git tool is: git
No credentials specified
Cloning the remote Git repository
Using shallow clone with depth 1
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_storage_test # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
 > git rev-list --no-walk 03312178c534dce949face80c69812d989e55009 # timeout=10
[Pipeline] }
[Pipeline] // stage
[Pipeline] withEnv
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 1 hr 0 min
[Pipeline] {
[Pipeline] stage
[Pipeline] { (Debug info)
[Pipeline] sh
+ printenv
PROW_JOB_ID=47a1001c-f251-49c4-8069-11a0a23be56a
JENKINS_NODE_COOKIE=801a5680-3cfe-41e2-a939-b6884a0fc7e1
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_storage_test/1763/
GOLANG_VERSION=1.21.6
HOSTNAME=pingcap-tiflow-pull-cdc-integration-storage-test-1763-42t-t5xkh
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT=tcp://10.233.0.1:443
KUBERNETES_PORT_443_TCP_PORT=443
TERM=xterm
STAGE_NAME=Debug info
BUILD_TAG=jenkins-pingcap-tiflow-pull_cdc_integration_storage_test-1763
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tiflow/pull_cdc_integration_storage_test","buildid":"1786618874523815938","prowjobid":"47a1001c-f251-49c4-8069-11a0a23be56a","refs":{"org":"pingcap","repo":"tiflow","repo_link":"https://github.com/pingcap/tiflow","base_ref":"master","base_sha":"be1553484fe4c03594eabb8d7435c694e5fd7224","base_link":"https://github.com/pingcap/tiflow/commit/be1553484fe4c03594eabb8d7435c694e5fd7224","pulls":[{"number":10919,"author":"lidezhu","sha":"b48aec50c527b43fe2205b46bcfc8c3dd5fe4448","title":"*(ticdc): split old update kv entry after restarting changefeed","link":"https://github.com/pingcap/tiflow/pull/10919","commit_link":"https://github.com/pingcap/tiflow/pull/10919/commits/b48aec50c527b43fe2205b46bcfc8c3dd5fe4448","author_link":"https://github.com/lidezhu"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_storage_test
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_storage_test/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_storage_test/1763/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_storage_test/1763/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_storage_test/1763/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
_=/usr/bin/printenv
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_storage_test
HUDSON_URL=https://do.pingcap.net/jenkins/
JOB_NAME=pingcap/tiflow/pull_cdc_integration_storage_test
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#1763
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1786618874523815938
GOLANG_DOWNLOAD_SHA256=3f934f40ac360b9c01f616a9aa1796d227d8b0328bf64cb045c7b8c4ee9caea4
JOB_BASE_NAME=pull_cdc_integration_storage_test
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_storage_test/1763/display/redirect?page=tests
SHLVL=3
HOME=/home/jenkins
POD_LABEL=pingcap_tiflow_pull_cdc_integration_storage_test_1763-42tt5
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_storage_test@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-ec41a9ea2975eb575c6c655de75271b1a69674254216af41486e1697505d3c9f
NODE_LABELS=pingcap-tiflow-pull-cdc-integration-storage-test-1763-42t-t5xkh pingcap_tiflow_pull_cdc_integration_storage_test_1763-42tt5
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tiflow-pull-cdc-integration-storage-test-1763-42t-t5xkh
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tiflow/job/pull_cdc_integration_storage_test/display/redirect
BUILD_NUMBER=1763
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.6.linux-amd64.tar.gz
+ echo -------------------------
-------------------------
+ go env
GO111MODULE=''
GOARCH='amd64'
GOBIN=''
GOCACHE='/home/jenkins/.cache/go-build'
GOENV='/home/jenkins/.config/go/env'
GOEXE=''
GOEXPERIMENT=''
GOFLAGS=''
GOHOSTARCH='amd64'
GOHOSTOS='linux'
GOINSECURE=''
GOMODCACHE='/go/pkg/mod'
GONOPROXY=''
GONOSUMDB=''
GOOS='linux'
GOPATH='/go'
GOPRIVATE=''
GOPROXY='http://goproxy.apps.svc,https://proxy.golang.org,direct'
GOROOT='/usr/local/go'
GOSUMDB='sum.golang.org'
GOTMPDIR=''
GOTOOLCHAIN='auto'
GOTOOLDIR='/usr/local/go/pkg/tool/linux_amd64'
GOVCS=''
GOVERSION='go1.21.6'
GCCGO='gccgo'
GOAMD64='v1'
AR='ar'
CC='gcc'
CXX='g++'
CGO_ENABLED='1'
GOMOD='/dev/null'
GOWORK=''
CGO_CFLAGS='-O2 -g'
CGO_CPPFLAGS=''
CGO_CXXFLAGS='-O2 -g'
CGO_FFLAGS='-O2 -g'
CGO_LDFLAGS='-O2 -g'
PKG_CONFIG='pkg-config'
GOGCCFLAGS='-fPIC -m64 -pthread -Wl,--no-gc-sections -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build2657570554=/tmp/go-build -gno-record-gcc-switches'
+ echo -------------------------
-------------------------
+ echo 'debug command: kubectl -n jenkins-tiflow exec -ti pingcap-tiflow-pull-cdc-integration-storage-test-1763-42t-t5xkh bash'
debug command: kubectl -n jenkins-tiflow exec -ti pingcap-tiflow-pull-cdc-integration-storage-test-1763-42t-t5xkh bash
[Pipeline] container
[Pipeline] {
[Pipeline] sh
+ dig github.com

; <<>> DiG 9.18.16 <<>> github.com
;; global options: +cmd
;; Got answer:
;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 36507
;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 1

;; OPT PSEUDOSECTION:
; EDNS: version: 0, flags:; udp: 1232
; COOKIE: 8634be3848916966 (echoed)
;; QUESTION SECTION:
;github.com.			IN	A

;; ANSWER SECTION:
github.com.		26	IN	A	20.205.243.166

;; Query time: 0 msec
;; SERVER: 169.254.25.10#53(169.254.25.10) (UDP)
;; WHEN: Sat May 04 04:48:53 UTC 2024
;; MSG SIZE  rcvd: 77

[Pipeline] script
[Pipeline] {
[Pipeline] }
[Pipeline] // script
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Check diff files)
[Pipeline] container
[Pipeline] {
[Pipeline] script
[Pipeline] {
[Pipeline] withCredentials
Masking supported pattern matches of $token
[Pipeline] {
[Pipeline] httpRequest
Warning: A secret was passed to "httpRequest" using Groovy String interpolation, which is insecure.
		 Affected argument(s) used the following variable(s): [token]
		 See https://jenkins.io/redirect/groovy-string-interpolation for details.
HttpMethod: GET
URL: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=1&per_page=100
Content-Type: application/json
Authorization: *****
Sending request to url: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=1&per_page=100
Response Code: HTTP/1.1 200 OK
Success: Status code 200 is in the accepted range: 100:399
[Pipeline] httpRequest
Warning: A secret was passed to "httpRequest" using Groovy String interpolation, which is insecure.
		 Affected argument(s) used the following variable(s): [token]
		 See https://jenkins.io/redirect/groovy-string-interpolation for details.
HttpMethod: GET
URL: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=2&per_page=100
Content-Type: application/json
Authorization: *****
Sending request to url: https://api.github.com/repos/pingcap/tiflow/pulls/10919/files?page=2&per_page=100
Response Code: HTTP/1.1 200 OK
Success: Status code 200 is in the accepted range: 100:399
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] echo
pr_diff_files: [cdc/model/kv.go, cdc/model/sink.go, cdc/model/sink_test.go, cdc/processor/processor.go, cdc/processor/sinkmanager/manager.go, cdc/processor/sourcemanager/manager.go, cdc/redo/reader/reader.go, cdc/sink/dmlsink/factory/factory.go, cdc/sink/dmlsink/txn/mysql/mysql.go, cdc/sink/dmlsink/txn/mysql/mysql_test.go, cmd/kafka-consumer/main.go, cmd/pulsar-consumer/main.go, cmd/storage-consumer/main.go, errors.toml, pkg/applier/redo.go, pkg/applier/redo_test.go, pkg/errors/cdc_errors.go, pkg/errors/helper.go, tests/integration_tests/changefeed_dup_error_restart/conf/diff_config.toml, tests/integration_tests/changefeed_dup_error_restart/conf/workload, tests/integration_tests/changefeed_dup_error_restart/run.sh, tests/integration_tests/force_replicate_table/run.sh, tests/integration_tests/open_protocol_handle_key_only/data/data.sql, tests/integration_tests/run_group.sh]
[Pipeline] echo
diff file not matched: cdc/model/kv.go
[Pipeline] }
[Pipeline] // script
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Checkout)
[Pipeline] timeout
Timeout set to expire in 10 min
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_storage_test/tiflow
[Pipeline] {
[Pipeline] cache
Cache restored successfully (git/pingcap/tiflow/rev-be15534)
203635712 bytes in 2.41 secs (84395342 bytes/sec)
[Pipeline] {
[Pipeline] retry
[Pipeline] {
[Pipeline] script
[Pipeline] {
[Pipeline] sh
git version 2.36.6
Reinitialized existing Git repository in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_storage_test/tiflow/.git/
.git
HEAD is now at be1553484 codec(ticdc): avro simplify the unit test (#11010)
POST git-upload-pack (656 bytes)
POST git-upload-pack (973 bytes)
From https://github.com/pingcap/tiflow
 = [up to date]          master               -> origin/master
 * [new ref]             refs/pull/10919/head -> origin/pr/10919/head
HEAD is now at be1553484 codec(ticdc): avro simplify the unit test (#11010)
๐Ÿšง Checkouting to base SHA:be1553484fe4c03594eabb8d7435c694e5fd7224...
HEAD is now at be1553484 codec(ticdc): avro simplify the unit test (#11010)
โœ… Checked. ๐ŸŽ‰
๐Ÿงพ HEAD info:
be1553484fe4c03594eabb8d7435c694e5fd7224
be1553484 codec(ticdc): avro simplify the unit test (#11010)
2a7a65c6f Support Sequences (#10203)
36e9e1bf6 cli(ticdc): allow client authentication to be enabled without tls (#11005)
๐Ÿšง Pre-merge heads of pull requests to base SHA: be1553484fe4c03594eabb8d7435c694e5fd7224 ...
Updating be1553484..b48aec50c
Fast-forward
 cdc/model/kv.go                                    |   5 +
 cdc/model/sink.go                                  |  35 ++-
 cdc/model/sink_test.go                             |   9 +-
 cdc/processor/processor.go                         |  21 +-
 cdc/processor/sinkmanager/manager.go               |   5 +
 cdc/processor/sourcemanager/manager.go             |  66 +++-
 cdc/redo/reader/reader.go                          |  21 +-
 cdc/sink/dmlsink/factory/factory.go                |   8 +-
 cdc/sink/dmlsink/txn/mysql/mysql.go                |  89 +++---
 cdc/sink/dmlsink/txn/mysql/mysql_test.go           |   2 +-
 cmd/kafka-consumer/main.go                         |   4 +-
 cmd/pulsar-consumer/main.go                        |  17 +-
 cmd/storage-consumer/main.go                       |   4 +-
 errors.toml                                        |   5 +
 pkg/applier/redo.go                                | 303 +++++++++++++++++-
 pkg/applier/redo_test.go                           | 347 ++++++++++++++++++++-
 pkg/errors/cdc_errors.go                           |   4 +
 pkg/errors/helper.go                               |  19 ++
 .../conf/diff_config.toml                          |  29 ++
 .../changefeed_dup_error_restart/conf/workload     |  13 +
 .../changefeed_dup_error_restart/run.sh            |  54 ++++
 .../integration_tests/force_replicate_table/run.sh |   4 +-
 .../open_protocol_handle_key_only/data/data.sql    |  71 +----
 tests/integration_tests/run_group.sh               |   5 +-
 24 files changed, 976 insertions(+), 164 deletions(-)
 create mode 100644 tests/integration_tests/changefeed_dup_error_restart/conf/diff_config.toml
 create mode 100644 tests/integration_tests/changefeed_dup_error_restart/conf/workload
 create mode 100755 tests/integration_tests/changefeed_dup_error_restart/run.sh
๐Ÿงพ Pre-merged result:
b48aec50c527b43fe2205b46bcfc8c3dd5fe4448
b48aec50c fix
54efd79f2 check
cafa249a0 check
โœ… Pre merged ๐ŸŽ‰
โœ… ~~~~~All done.~~~~~~
[Pipeline] }
[Pipeline] // script
[Pipeline] }
[Pipeline] // retry
[Pipeline] }
Cache saved successfully (git/pingcap/tiflow/rev-be15534-b48aec5)
203831296 bytes in 4.25 secs (47988402 bytes/sec)
[Pipeline] // cache
[Pipeline] }
[Pipeline] // dir
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (prepare)
[Pipeline] timeout
Timeout set to expire in 20 min
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tiflow/pull_cdc_integration_storage_test/third_party_download
[Pipeline] {
[Pipeline] retry
[Pipeline] {
[Pipeline] sh
+ cd ../tiflow
+ ./scripts/download-integration-test-binaries.sh master
Download binaries...
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0    534      0 --:--:-- --:--:-- --:--:--   539
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0   2160      0 --:--:-- --:--:-- --:--:--  2277
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0   1684      0 --:--:-- --:--:-- --:--:--  1708
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100    41  100    41    0     0    362      0 --:--:-- --:--:-- --:--:--   362
>>>
download tidb-server.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/tidb/600b2ed4bf0aa38224a1c4c4c68831820735515c/centos7/tidb-server.tar.gz
2024-05-04 12:49:17 URL:http://fileserver.pingcap.net/download/builds/pingcap/tidb/600b2ed4bf0aa38224a1c4c4c68831820735515c/centos7/tidb-server.tar.gz [536570515/536570515] -> "tmp/tidb-server.tar.gz" [1]
>>>
download pd-server.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/pd/1679dbca25b3483d1375c7e747da27e99ad77360/centos7/pd-server.tar.gz
2024-05-04 12:49:29 URL:http://fileserver.pingcap.net/download/builds/pingcap/pd/1679dbca25b3483d1375c7e747da27e99ad77360/centos7/pd-server.tar.gz [187372022/187372022] -> "tmp/pd-server.tar.gz" [1]
>>>
download tikv-server.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/tikv/72a0fd5b00235a7c56014b77ddd933e2a0d33c88/centos7/tikv-server.tar.gz
2024-05-04 12:49:43 URL:http://fileserver.pingcap.net/download/builds/pingcap/tikv/72a0fd5b00235a7c56014b77ddd933e2a0d33c88/centos7/tikv-server.tar.gz [919098782/919098782] -> "tmp/tikv-server.tar.gz" [1]
>>>
download tiflash.tar.gz from http://fileserver.pingcap.net/download/builds/pingcap/tiflash/master/8e170090fad91c94bef8d908e21c195c1d145b02/centos7/tiflash.tar.gz
Aborted by Jenkins Admin
Sending interrupt signal to process
Killing processes
kill finished with exit code 0
script returned exit code 143
[Pipeline] }
[Pipeline] // retry
[Pipeline] }
[Pipeline] // dir
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Tests)
Stage "Tests" skipped due to earlier failure(s)
[Pipeline] parallel
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G00')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G01')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G02')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G03')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G04')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G05')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G06')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G07')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G08')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G09')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G10')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G11')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G12')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G13')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G14')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G15')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G16')
[Pipeline] { (Branch: Matrix - TEST_GROUP = 'G17')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G00')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G01')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G02')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G03')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G04')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G05')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G06')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G07')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G08')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G09')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G10')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G11')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G12')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G13')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G14')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G15')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G16')
[Pipeline] stage
[Pipeline] { (Matrix - TEST_GROUP = 'G17')
Stage "Matrix - TEST_GROUP = 'G00'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G01'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G02'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G03'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G04'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G05'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G06'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G07'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G08'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G09'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G10'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G11'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G12'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G13'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G14'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G15'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G16'" skipped due to earlier failure(s)
Stage "Matrix - TEST_GROUP = 'G17'" skipped due to earlier failure(s)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
Stage "Matrix - TEST_GROUP = 'G00'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G01'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G02'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G03'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G04'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G05'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G06'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G07'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G08'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G09'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G10'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G11'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G12'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G13'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G14'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G15'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G16'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - TEST_GROUP = 'G17'" skipped due to earlier failure(s)
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G00'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G01'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G02'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G03'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G04'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G05'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G06'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G07'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G08'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G09'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G10'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G11'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G12'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G13'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G14'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G15'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G16'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G17'
[Pipeline] // parallel
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] End of Pipeline
org.jenkinsci.plugins.workflow.actions.ErrorAction$ErrorId: 61558f3e-6b2c-4202-811b-8eda405ab2ce
Finished: ABORTED