Skip to content

Console Output

Started by user Jenkins Admin
Obtained pipelines/pingcap/tidb/release-8.1/ghpr_check2.groovy from git https://github.com/PingCAP-QE/ci.git
Loading library tipipeline@main
Library tipipeline@main is cached. Copying from home.
[Pipeline] Start of Pipeline
[Pipeline] readJSON
[Pipeline] readTrusted
Obtained pipelines/pingcap/tidb/release-8.1/pod-ghpr_check2.yaml from git https://github.com/PingCAP-QE/ci.git
[Pipeline] podTemplate
[Pipeline] {
[Pipeline] node
Created Pod: kubernetes jenkins-tidb/pingcap-tidb-release-8-1-ghpr-check2-112-htdxc-v9cr7-sq06b
Agent pingcap-tidb-release-8-1-ghpr-check2-112-htdxc-v9cr7-sq06b is provisioned from template pingcap_tidb_release-8_1_ghpr_check2_112-htdxc-v9cr7
---
apiVersion: "v1"
kind: "Pod"
metadata:
  annotations:
    buildUrl: "http://jenkins.apps.svc.cluster.local:8080/jenkins/job/pingcap/job/tidb/job/release-8.1/job/ghpr_check2/112/"
    runUrl: "job/pingcap/job/tidb/job/release-8.1/job/ghpr_check2/112/"
  labels:
    jenkins/jenkins-jenkins-agent: "true"
    jenkins/label-digest: "f6f43ac52a7031bf489cb1ac7ec198ad436a2c72"
    jenkins/label: "pingcap_tidb_release-8_1_ghpr_check2_112-htdxc"
  name: "pingcap-tidb-release-8-1-ghpr-check2-112-htdxc-v9cr7-sq06b"
  namespace: "jenkins-tidb"
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: "kubernetes.io/arch"
            operator: "In"
            values:
            - "amd64"
          - key: "ci-nvme-high-performance"
            operator: "In"
            values:
            - "true"
  containers:
  - image: "hub.pingcap.net/wangweizhen/tidb_image:go12120230809"
    lifecycle:
      postStart:
        exec:
          command:
          - "/bin/sh"
          - "/data/bazel-prepare-in-container.sh"
    name: "golang"
    resources:
      limits:
        memory: "32Gi"
        cpu: "8"
    securityContext:
      privileged: true
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/.tidb/tmp"
      name: "bazel-out-merged"
    - mountPath: "/bazel-out-lower"
      name: "bazel-out-lower"
      subPath: "tidb/go1.19.2"
    - mountPath: "/bazel-out-overlay"
      name: "bazel-out-overlay"
    - mountPath: "/share/.cache/go-build"
      name: "gocache"
    - mountPath: "/share/.go"
      name: "gopathcache"
    - mountPath: "/share/.cache/bazel-repository-cache"
      name: "bazel-repository-cache"
    - mountPath: "/data/"
      name: "bazel-rc"
      readOnly: true
    - mountPath: "/etc/containerinfo"
      name: "containerinfo"
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/network-multitool"
    name: "net-tool"
    resources:
      limits:
        memory: "128Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - image: "hub.pingcap.net/jenkins/python3-requests:latest"
    name: "report"
    resources:
      limits:
        memory: "256Mi"
        cpu: "100m"
    tty: true
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  - env:
    - name: "JENKINS_SECRET"
      value: "********"
    - name: "JENKINS_TUNNEL"
      value: "jenkins-agent.apps.svc.cluster.local:50000"
    - name: "JENKINS_AGENT_NAME"
      value: "pingcap-tidb-release-8-1-ghpr-check2-112-htdxc-v9cr7-sq06b"
    - name: "JENKINS_NAME"
      value: "pingcap-tidb-release-8-1-ghpr-check2-112-htdxc-v9cr7-sq06b"
    - name: "JENKINS_AGENT_WORKDIR"
      value: "/home/jenkins/agent"
    - name: "JENKINS_URL"
      value: "http://jenkins.apps.svc.cluster.local:8080/jenkins/"
    image: "jenkins/inbound-agent:3206.vb_15dcf73f6a_9-2"
    name: "jnlp"
    resources:
      requests:
        memory: "256Mi"
        cpu: "100m"
    volumeMounts:
    - mountPath: "/home/jenkins/agent"
      name: "workspace-volume"
      readOnly: false
  restartPolicy: "Never"
  securityContext:
    fsGroup: 1000
  volumes:
  - emptyDir: {}
    name: "bazel-out-merged"
  - name: "bazel-rc"
    secret:
      secretName: "bazel"
  - emptyDir: {}
    name: "bazel-out-overlay"
  - name: "gocache"
    persistentVolumeClaim:
      claimName: "gocache"
  - downwardAPI:
      items:
      - path: "cpu_limit"
        resourceFieldRef:
          containerName: "golang"
          resource: "limits.cpu"
      - path: "cpu_request"
        resourceFieldRef:
          containerName: "golang"
          resource: "requests.cpu"
      - path: "mem_limit"
        resourceFieldRef:
          containerName: "golang"
          resource: "limits.memory"
      - path: "mem_request"
        resourceFieldRef:
          containerName: "golang"
          resource: "requests.memory"
    name: "containerinfo"
  - name: "bazel-out-lower"
    persistentVolumeClaim:
      claimName: "bazel-out-data"
  - emptyDir:
      medium: ""
    name: "workspace-volume"
  - name: "gopathcache"
    persistentVolumeClaim:
      claimName: "gopathcache"
  - name: "bazel-repository-cache"
    persistentVolumeClaim:
      claimName: "bazel-repository-cache"

Running on pingcap-tidb-release-8-1-ghpr-check2-112-htdxc-v9cr7-sq06b in /home/jenkins/agent/workspace/pingcap/tidb/release-8.1/ghpr_check2
[Pipeline] {
[Pipeline] stage
[Pipeline] { (Declarative: Checkout SCM)
[Pipeline] checkout
The recommended git tool is: git
No credentials specified
Cloning the remote Git repository
Using shallow clone with depth 1
Cloning repository https://github.com/PingCAP-QE/ci.git
 > git init /home/jenkins/agent/workspace/pingcap/tidb/release-8.1/ghpr_check2 # timeout=10
Fetching upstream changes from https://github.com/PingCAP-QE/ci.git
 > git --version # timeout=10
 > git --version # 'git version 2.39.2'
 > git fetch --tags --force --progress --depth=1 -- https://github.com/PingCAP-QE/ci.git +refs/heads/*:refs/remotes/origin/* # timeout=5
 > git config remote.origin.url https://github.com/PingCAP-QE/ci.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
Avoid second fetch
Checking out Revision 03312178c534dce949face80c69812d989e55009 (origin/main)
Commit message: "fix(br): use failpoint tidb-server instead (#2951)"
 > git rev-parse origin/main^{commit} # timeout=10
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 03312178c534dce949face80c69812d989e55009 # timeout=10
 > git rev-list --no-walk 03312178c534dce949face80c69812d989e55009 # timeout=10
[Pipeline] }
[Pipeline] // stage
[Pipeline] withEnv
[Pipeline] {
[Pipeline] container
[Pipeline] {
[Pipeline] withEnv
[Pipeline] {
[Pipeline] timeout
Timeout set to expire in 1 hr 5 min
[Pipeline] {
[Pipeline] stage
[Pipeline] { (Debug info)
[Pipeline] sh
+ printenv
PROW_JOB_ID=efa5b717-65cd-484c-b11a-a174b0b68da6
JENKINS_NODE_COOKIE=15aed13a-9824-4a39-bd63-6acbb42e5514
BUILD_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tidb/job/release-8.1/job/ghpr_check2/112/
GOLANG_VERSION=1.21.0
HOSTNAME=pingcap-tidb-release-8-1-ghpr-check2-112-htdxc-v9cr7-sq06b
HUDSON_SERVER_COOKIE=83ef27fe9acccc92
KUBERNETES_PORT_443_TCP_PORT=443
KUBERNETES_PORT=tcp://10.233.0.1:443
TERM=xterm
STAGE_NAME=Debug info
BUILD_TAG=jenkins-pingcap-tidb-release-8.1-ghpr_check2-112
KUBERNETES_SERVICE_PORT=443
GIT_PREVIOUS_COMMIT=03312178c534dce949face80c69812d989e55009
JOB_SPEC={"type":"presubmit","job":"pingcap/tidb/release-8.1/ghpr_check2","buildid":"1787323754439774211","prowjobid":"efa5b717-65cd-484c-b11a-a174b0b68da6","refs":{"org":"pingcap","repo":"tidb","repo_link":"https://github.com/pingcap/tidb","base_ref":"release-8.1","base_sha":"a088e599ec5e23c52f53da21fa9245780dcc6331","base_link":"https://github.com/pingcap/tidb/commit/a088e599ec5e23c52f53da21fa9245780dcc6331","pulls":[{"number":53025,"author":"zimulala","sha":"fd7df399ece6a0cb02dd94066fd8692c7e94c49b","title":"tests/ddl: make default_as_expression test stable","link":"https://github.com/pingcap/tidb/pull/53025","commit_link":"https://github.com/pingcap/tidb/pull/53025/commits/fd7df399ece6a0cb02dd94066fd8692c7e94c49b","author_link":"https://github.com/zimulala"}]}}
KUBERNETES_SERVICE_HOST=10.233.0.1
WORKSPACE=/home/jenkins/agent/workspace/pingcap/tidb/release-8.1/ghpr_check2
JOB_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tidb/job/release-8.1/job/ghpr_check2/
RUN_CHANGES_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tidb/job/release-8.1/job/ghpr_check2/112/display/redirect?page=changes
RUN_ARTIFACTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tidb/job/release-8.1/job/ghpr_check2/112/display/redirect?page=artifacts
FILE_SERVER_URL=http://fileserver.pingcap.net
JENKINS_HOME=/var/jenkins_home
GIT_COMMIT=03312178c534dce949face80c69812d989e55009
PATH=/go/bin:/usr/local/go/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
RUN_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tidb/job/release-8.1/job/ghpr_check2/112/display/redirect
GOPROXY=http://goproxy.apps.svc,https://proxy.golang.org,direct
_=/usr/bin/printenv
POD_CONTAINER=golang
PWD=/home/jenkins/agent/workspace/pingcap/tidb/release-8.1/ghpr_check2
HUDSON_URL=https://do.pingcap.net/jenkins/
JOB_NAME=pingcap/tidb/release-8.1/ghpr_check2
TZ=Asia/Shanghai
BUILD_DISPLAY_NAME=#112
JENKINS_URL=https://do.pingcap.net/jenkins/
BUILD_ID=1787323754439774211
GOLANG_DOWNLOAD_SHA256=d0398903a16ba2232b389fb31032ddf57cac34efda306a0eebac34f0965a0742
JOB_BASE_NAME=ghpr_check2
GIT_PREVIOUS_SUCCESSFUL_COMMIT=03312178c534dce949face80c69812d989e55009
RUN_TESTS_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tidb/job/release-8.1/job/ghpr_check2/112/display/redirect?page=tests
SHLVL=3
HOME=/home/jenkins
POD_LABEL=pingcap_tidb_release-8_1_ghpr_check2_112-htdxc
GOROOT=/usr/local/go
GIT_BRANCH=origin/main
KUBERNETES_PORT_443_TCP_PROTO=tcp
CI=true
KUBERNETES_SERVICE_PORT_HTTPS=443
WORKSPACE_TMP=/home/jenkins/agent/workspace/pingcap/tidb/release-8.1/ghpr_check2@tmp
EXECUTOR_NUMBER=0
JENKINS_SERVER_COOKIE=durable-7f8296a37cf59b5fc344e1c2a8f416fdcd8370c64af98c32b8ac1b197024be9d
NODE_LABELS=pingcap_tidb_release-8_1_ghpr_check2_112-htdxc pingcap-tidb-release-8-1-ghpr-check2-112-htdxc-v9cr7-sq06b
GIT_URL=https://github.com/PingCAP-QE/ci.git
HUDSON_HOME=/var/jenkins_home
CLASSPATH=
NODE_NAME=pingcap-tidb-release-8-1-ghpr-check2-112-htdxc-v9cr7-sq06b
GOPATH=/go
JOB_DISPLAY_URL=https://do.pingcap.net/jenkins/job/pingcap/job/tidb/job/release-8.1/job/ghpr_check2/display/redirect
BUILD_NUMBER=112
KUBERNETES_PORT_443_TCP_ADDR=10.233.0.1
KUBERNETES_PORT_443_TCP=tcp://10.233.0.1:443
GOLANG_DOWNLOAD_URL=https://dl.google.com/go/go1.21.0.linux-amd64.tar.gz
+ echo -------------------------
-------------------------
+ go env
GO111MODULE=''
GOARCH='amd64'
GOBIN=''
GOCACHE='/home/jenkins/.cache/go-build'
GOENV='/home/jenkins/.config/go/env'
GOEXE=''
GOEXPERIMENT=''
GOFLAGS=''
GOHOSTARCH='amd64'
GOHOSTOS='linux'
GOINSECURE=''
GOMODCACHE='/go/pkg/mod'
GONOPROXY=''
GONOSUMDB=''
GOOS='linux'
GOPATH='/go'
GOPRIVATE=''
GOPROXY='http://goproxy.apps.svc,https://proxy.golang.org,direct'
GOROOT='/usr/local/go'
GOSUMDB='sum.golang.org'
GOTMPDIR=''
GOTOOLCHAIN='auto'
GOTOOLDIR='/usr/local/go/pkg/tool/linux_amd64'
GOVCS=''
GOVERSION='go1.21.0'
GCCGO='gccgo'
GOAMD64='v1'
AR='ar'
CC='gcc'
CXX='g++'
CGO_ENABLED='1'
GOMOD='/dev/null'
GOWORK=''
CGO_CFLAGS='-O2 -g'
CGO_CPPFLAGS=''
CGO_CXXFLAGS='-O2 -g'
CGO_FFLAGS='-O2 -g'
CGO_LDFLAGS='-O2 -g'
PKG_CONFIG='pkg-config'
GOGCCFLAGS='-fPIC -m64 -pthread -Wl,--no-gc-sections -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build3962721428=/tmp/go-build -gno-record-gcc-switches'
+ echo -------------------------
-------------------------
+ echo 'debug command: kubectl -n jenkins-tidb exec -ti pingcap-tidb-release-8-1-ghpr-check2-112-htdxc-v9cr7-sq06b bash'
debug command: kubectl -n jenkins-tidb exec -ti pingcap-tidb-release-8-1-ghpr-check2-112-htdxc-v9cr7-sq06b bash
[Pipeline] container
[Pipeline] {
[Pipeline] sh
+ dig github.com

; <<>> DiG 9.18.16 <<>> github.com
;; global options: +cmd
;; Got answer:
;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 63971
;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 1

;; OPT PSEUDOSECTION:
; EDNS: version: 0, flags:; udp: 1232
; COOKIE: b87af2f1af301d42 (echoed)
;; QUESTION SECTION:
;github.com.			IN	A

;; ANSWER SECTION:
github.com.		24	IN	A	20.205.243.166

;; Query time: 0 msec
;; SERVER: 169.254.25.10#53(169.254.25.10) (UDP)
;; WHEN: Mon May 06 03:29:59 UTC 2024
;; MSG SIZE  rcvd: 77

[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Checkout)
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/release-8.1/ghpr_check2/tidb
[Pipeline] {
[Pipeline] cache
Cache restored successfully (git/pingcap/tidb/rev-a088e59-c4be2c0)
459142656 bytes in 3.21 secs (142865606 bytes/sec)
[Pipeline] {
[Pipeline] retry
[Pipeline] {
[Pipeline] script
[Pipeline] {
[Pipeline] sh
git version 2.37.2
Reinitialized existing Git repository in /home/jenkins/agent/workspace/pingcap/tidb/release-8.1/ghpr_check2/tidb/.git/
.git
HEAD is now at c4be2c041c done
POST git-upload-pack (686 bytes)
POST git-upload-pack (973 bytes)
POST git-upload-pack (gzip 2023 to 1074 bytes)
From https://github.com/pingcap/tidb
 = [up to date]            release-8.1          -> origin/release-8.1
 * [new ref]               refs/pull/53025/head -> origin/pr/53025/head
Previous HEAD position was c4be2c041c done
HEAD is now at a088e599ec *: bump client-go to fix #53011 | tidb-test=pr/2320 (#53012) (#53015)
🚧 Checkouting to base SHA:a088e599ec5e23c52f53da21fa9245780dcc6331...
HEAD is now at a088e599ec *: bump client-go to fix #53011 | tidb-test=pr/2320 (#53012) (#53015)
✅ Checked. 🎉
🧾 HEAD info:
a088e599ec5e23c52f53da21fa9245780dcc6331
a088e599ec *: bump client-go to fix #53011 | tidb-test=pr/2320 (#53012) (#53015)
4400134511 infoschema: load auto id related changes for multi-schema-change diff | tidb-test=pr/2320 (#52967) (#53008)
9930a02ce0 ddl: fix the issue about create columns with wrong default when sql_mode is empty. (#52988) (#53002)
🚧 Pre-merge heads of pull requests to base SHA: a088e599ec5e23c52f53da21fa9245780dcc6331 ...
Auto-merging tests/integrationtest/r/ddl/default_as_expression.result
CONFLICT (content): Merge conflict in tests/integrationtest/r/ddl/default_as_expression.result
Automatic merge failed; fix conflicts and then commit the result.
[Pipeline] }
[Pipeline] // script
[Pipeline] }
ERROR: script returned exit code 1
Retrying
[Pipeline] {
[Pipeline] script
[Pipeline] {
[Pipeline] sh
git version 2.37.2
Reinitialized existing Git repository in /home/jenkins/agent/workspace/pingcap/tidb/release-8.1/ghpr_check2/tidb/.git/
.git
HEAD is now at a088e599ec *: bump client-go to fix #53011 | tidb-test=pr/2320 (#53012) (#53015)
POST git-upload-pack (686 bytes)
From https://github.com/pingcap/tidb
 = [up to date]            release-8.1          -> origin/release-8.1
 = [up to date]            refs/pull/53025/head -> origin/pr/53025/head
HEAD is now at a088e599ec *: bump client-go to fix #53011 | tidb-test=pr/2320 (#53012) (#53015)
🚧 Checkouting to base SHA:a088e599ec5e23c52f53da21fa9245780dcc6331...
HEAD is now at a088e599ec *: bump client-go to fix #53011 | tidb-test=pr/2320 (#53012) (#53015)
✅ Checked. 🎉
🧾 HEAD info:
a088e599ec5e23c52f53da21fa9245780dcc6331
a088e599ec *: bump client-go to fix #53011 | tidb-test=pr/2320 (#53012) (#53015)
4400134511 infoschema: load auto id related changes for multi-schema-change diff | tidb-test=pr/2320 (#52967) (#53008)
9930a02ce0 ddl: fix the issue about create columns with wrong default when sql_mode is empty. (#52988) (#53002)
🚧 Pre-merge heads of pull requests to base SHA: a088e599ec5e23c52f53da21fa9245780dcc6331 ...
Auto-merging tests/integrationtest/r/ddl/default_as_expression.result
CONFLICT (content): Merge conflict in tests/integrationtest/r/ddl/default_as_expression.result
Automatic merge failed; fix conflicts and then commit the result.
[Pipeline] }
[Pipeline] // script
[Pipeline] }
[Pipeline] // retry
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] // cache
[Pipeline] }
[Pipeline] // dir
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Prepare)
Stage "Prepare" skipped due to earlier failure(s)
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Checks)
Stage "Checks" skipped due to earlier failure(s)
[Pipeline] parallel
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh y')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh n')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_brietest')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pessimistictest')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_sessiontest')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_statisticstest')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_txntest')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest1')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest2')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest3')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest4')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest2')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest3')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest4')
[Pipeline] { (Branch: Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pipelineddmltest')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh y')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh n')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_brietest')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pessimistictest')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_sessiontest')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_statisticstest')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_txntest')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest1')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest2')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest3')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest4')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest2')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest3')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest4')
[Pipeline] stage
[Pipeline] { (Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pipelineddmltest')
Stage "Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh y'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh n'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_brietest'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pessimistictest'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_sessiontest'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_statisticstest'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_txntest'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest1'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest2'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest3'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest4'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest2'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest3'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest4'" skipped due to earlier failure(s)
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pipelineddmltest'" skipped due to earlier failure(s)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
[Pipeline] stage
[Pipeline] { (Test)
Stage "Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh y'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh n'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_brietest'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pessimistictest'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_sessiontest'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_statisticstest'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_txntest'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest1'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest2'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest3'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest4'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest2'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest3'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest4'" skipped due to earlier failure(s)
[Pipeline] }
Stage "Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pipelineddmltest'" skipped due to earlier failure(s)
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh y'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh n'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_brietest'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pessimistictest'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_sessiontest'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_statisticstest'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_txntest'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest1'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest2'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest3'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest4'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest2'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest3'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest4'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pipelineddmltest'
[Pipeline] // parallel
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Declarative: Post Actions)
[Pipeline] container
[Pipeline] {
[Pipeline] sh
+ bash scripts/plugins/report_job_result.sh FAILURE result.json
http://fileserver.pingcap.net
--2024-05-06 03:30:22--  http://fileserver.pingcap.net/download/rd-atom-agent/agent_upload_verifyci_metadata.py
Resolving fileserver.pingcap.net (fileserver.pingcap.net)... 10.2.12.82
Connecting to fileserver.pingcap.net (fileserver.pingcap.net)|10.2.12.82|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 4181 (4.1K) [application/octet-stream]
Saving to: ‘agent_upload_verifyci_metadata.py’

     0K ....                                                  100%  438M=0s

2024-05-06 03:30:22 (438 MB/s) - ‘agent_upload_verifyci_metadata.py’ saved [4181/4181]

No junit report file
parse result file result.json success
upload data succesfully.
[Pipeline] }
[Pipeline] // container
[Pipeline] archiveArtifacts
Archiving artifacts
Recording fingerprints
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] End of Pipeline
ERROR: script returned exit code 1
Finished: FAILURE