Skip to content

Console Output

Skipping 444 KB.. Full Log
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
+ git status
HEAD detached from origin/master
You are in a sparse checkout with 100% of tracked files present.

Changes not staged for commit:
  (use "git add <file>..." to update what will be committed)
  (use "git restore <file>..." to discard changes in working directory)
	modified:   Makefile.common

Untracked files:
  (use "git add <file>..." to include in what will be committed)
	rev-a3d52d59f14ed8131b5b84f8e3876fc0c32f1ffa

no changes added to commit (use "git add" and/or "git commit -a")
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] // timeout
[Pipeline] }
[Pipeline] sh
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] sh
+ str='run_real_tikv_tests.sh bazel_addindextest3'
+ logs_dir=logs_run_real_tikv_tests.sh_bazel_addindextest3
+ mkdir -p logs_run_real_tikv_tests.sh_bazel_addindextest3
+ mv 'pd*.log' logs_run_real_tikv_tests.sh_bazel_addindextest3
mv: cannot stat 'pd*.log': No such file or directory
+ true
+ mv 'tikv*.log' logs_run_real_tikv_tests.sh_bazel_addindextest3
mv: cannot stat 'tikv*.log': No such file or directory
+ true
+ mv tests/integrationtest/integration-test.out logs_run_real_tikv_tests.sh_bazel_addindextest3
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_run_real_tikv_tests.sh_bazel_addindextest3.tar.gz logs_run_real_tikv_tests.sh_bazel_addindextest3
logs_run_real_tikv_tests.sh_bazel_addindextest3/
[Pipeline] sh
+ git status
HEAD detached from origin/master
You are in a sparse checkout with 100% of tracked files present.

Changes not staged for commit:
  (use "git add <file>..." to update what will be committed)
  (use "git restore <file>..." to discard changes in working directory)
	modified:   Makefile.common

Untracked files:
  (use "git add <file>..." to include in what will be committed)
	rev-a3d52d59f14ed8131b5b84f8e3876fc0c32f1ffa

no changes added to commit (use "git add" and/or "git commit -a")
+ str='run_real_tikv_tests.sh bazel_pessimistictest'
+ logs_dir=logs_run_real_tikv_tests.sh_bazel_pessimistictest
+ mkdir -p logs_run_real_tikv_tests.sh_bazel_pessimistictest
+ mv 'pd*.log' logs_run_real_tikv_tests.sh_bazel_pessimistictest
mv: cannot stat 'pd*.log': No such file or directory
+ true
+ mv 'tikv*.log' logs_run_real_tikv_tests.sh_bazel_pessimistictest
mv: cannot stat 'tikv*.log': No such file or directory
+ true
+ mv tests/integrationtest/integration-test.out logs_run_real_tikv_tests.sh_bazel_pessimistictest
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_run_real_tikv_tests.sh_bazel_pessimistictest.tar.gz logs_run_real_tikv_tests.sh_bazel_pessimistictest
logs_run_real_tikv_tests.sh_bazel_pessimistictest/
[Pipeline] sh
+ git status
HEAD detached from origin/master
You are in a sparse checkout with 100% of tracked files present.

Changes not staged for commit:
  (use "git add <file>..." to update what will be committed)
  (use "git restore <file>..." to discard changes in working directory)
	modified:   Makefile.common

Untracked files:
  (use "git add <file>..." to include in what will be committed)
	rev-a3d52d59f14ed8131b5b84f8e3876fc0c32f1ffa

no changes added to commit (use "git add" and/or "git commit -a")
+ str='run_real_tikv_tests.sh bazel_brietest'
+ logs_dir=logs_run_real_tikv_tests.sh_bazel_brietest
+ mkdir -p logs_run_real_tikv_tests.sh_bazel_brietest
+ mv 'pd*.log' logs_run_real_tikv_tests.sh_bazel_brietest
mv: cannot stat 'pd*.log': No such file or directory
+ true
+ mv 'tikv*.log' logs_run_real_tikv_tests.sh_bazel_brietest
mv: cannot stat 'tikv*.log': No such file or directory
+ true
+ mv tests/integrationtest/integration-test.out logs_run_real_tikv_tests.sh_bazel_brietest
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_run_real_tikv_tests.sh_bazel_brietest.tar.gz logs_run_real_tikv_tests.sh_bazel_brietest
logs_run_real_tikv_tests.sh_bazel_brietest/
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] sh
+ str='run_real_tikv_tests.sh bazel_statisticstest'
+ logs_dir=logs_run_real_tikv_tests.sh_bazel_statisticstest
+ mkdir -p logs_run_real_tikv_tests.sh_bazel_statisticstest
+ mv 'pd*.log' logs_run_real_tikv_tests.sh_bazel_statisticstest
mv: cannot stat 'pd*.log': No such file or directory
+ true
+ mv 'tikv*.log' logs_run_real_tikv_tests.sh_bazel_statisticstest
mv: cannot stat 'tikv*.log': No such file or directory
+ true
+ mv tests/integrationtest/integration-test.out logs_run_real_tikv_tests.sh_bazel_statisticstest
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_run_real_tikv_tests.sh_bazel_statisticstest.tar.gz logs_run_real_tikv_tests.sh_bazel_statisticstest
logs_run_real_tikv_tests.sh_bazel_statisticstest/
[Pipeline] sh
GOBIN=/home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb/tools/bin GO111MODULE=on go install github.com/pingcap/failpoint/failpoint-ctl@2eaa328
go: downloading github.com/pingcap/failpoint v0.0.0-20220801062533-2eaa32854a6c
go: downloading github.com/sergi/go-diff v1.1.0
[Pipeline] sh
+ str='run_real_tikv_tests.sh bazel_txntest'
+ logs_dir=logs_run_real_tikv_tests.sh_bazel_txntest
+ mkdir -p logs_run_real_tikv_tests.sh_bazel_txntest
+ mv pd1.log pd2.log pd3.log logs_run_real_tikv_tests.sh_bazel_txntest
+ mv tikv1.log tikv2.log tikv3.log logs_run_real_tikv_tests.sh_bazel_txntest
+ mv tests/integrationtest/integration-test.out logs_run_real_tikv_tests.sh_bazel_txntest
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_run_real_tikv_tests.sh_bazel_txntest.tar.gz logs_run_real_tikv_tests.sh_bazel_txntest
logs_run_real_tikv_tests.sh_bazel_txntest/
logs_run_real_tikv_tests.sh_bazel_txntest/tikv3.log
logs_run_real_tikv_tests.sh_bazel_txntest/pd1.log
logs_run_real_tikv_tests.sh_bazel_txntest/tikv2.log
logs_run_real_tikv_tests.sh_bazel_txntest/pd3.log
logs_run_real_tikv_tests.sh_bazel_txntest/pd2.log
logs_run_real_tikv_tests.sh_bazel_txntest/tikv1.log
[Pipeline] sh
+ str='run_real_tikv_tests.sh bazel_addindextest'
+ logs_dir=logs_run_real_tikv_tests.sh_bazel_addindextest
+ mkdir -p logs_run_real_tikv_tests.sh_bazel_addindextest
+ mv pd1.log pd2.log pd3.log logs_run_real_tikv_tests.sh_bazel_addindextest
+ mv tikv1.log tikv2.log tikv3.log logs_run_real_tikv_tests.sh_bazel_addindextest
+ mv tests/integrationtest/integration-test.out logs_run_real_tikv_tests.sh_bazel_addindextest
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_run_real_tikv_tests.sh_bazel_addindextest.tar.gz logs_run_real_tikv_tests.sh_bazel_addindextest
logs_run_real_tikv_tests.sh_bazel_addindextest/
logs_run_real_tikv_tests.sh_bazel_addindextest/tikv1.log
logs_run_real_tikv_tests.sh_bazel_addindextest/pd1.log
logs_run_real_tikv_tests.sh_bazel_addindextest/tikv2.log
logs_run_real_tikv_tests.sh_bazel_addindextest/pd2.log
logs_run_real_tikv_tests.sh_bazel_addindextest/tikv3.log
logs_run_real_tikv_tests.sh_bazel_addindextest/pd3.log
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
+ str='run_real_tikv_tests.sh bazel_addindextest1'
+ logs_dir=logs_run_real_tikv_tests.sh_bazel_addindextest1
+ mkdir -p logs_run_real_tikv_tests.sh_bazel_addindextest1
+ mv pd1.log pd2.log pd3.log logs_run_real_tikv_tests.sh_bazel_addindextest1
+ mv tikv1.log tikv2.log tikv3.log logs_run_real_tikv_tests.sh_bazel_addindextest1
+ mv tests/integrationtest/integration-test.out logs_run_real_tikv_tests.sh_bazel_addindextest1
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_run_real_tikv_tests.sh_bazel_addindextest1.tar.gz logs_run_real_tikv_tests.sh_bazel_addindextest1
logs_run_real_tikv_tests.sh_bazel_addindextest1/
logs_run_real_tikv_tests.sh_bazel_addindextest1/tikv1.log
logs_run_real_tikv_tests.sh_bazel_addindextest1/pd1.log
logs_run_real_tikv_tests.sh_bazel_addindextest1/pd2.log
logs_run_real_tikv_tests.sh_bazel_addindextest1/tikv3.log
logs_run_real_tikv_tests.sh_bazel_addindextest1/pd3.log
logs_run_real_tikv_tests.sh_bazel_addindextest1/tikv2.log
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest4'
[Pipeline] // stage
[Pipeline] }
+ /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/scripts/pingcap/tidb/integrationtest_with_tikv.sh n
~/agent/workspace/pingcap/tidb/ghpr_check2/tidb/tests/integrationtest ~/agent/workspace/pingcap/tidb/ghpr_check2/tidb
extracting statistics: s
[Pipeline] sh
skip building tidb-server, using existing binary: /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb/bin/integration_test_tidb-server
building portgenerator binary: ./portgenerator
go: downloading go.uber.org/zap v1.27.0
go: downloading github.com/pingcap/log v1.1.1-0.20240314023424-862ccc32f18d
go: downloading github.com/phayes/freeport v0.0.0-20180830031419-95f893ade6f2
go: downloading gopkg.in/natefinch/lumberjack.v2 v2.2.1
go: downloading go.uber.org/multierr v1.11.0
[Pipeline] sh
+ /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/scripts/pingcap/tidb/run_real_tikv_tests.sh bazel_flashbacktest
[Pipeline] archiveArtifacts
Archiving artifacts
[Pipeline] archiveArtifacts
Archiving artifacts
[Pipeline] archiveArtifacts
Archiving artifacts
+ /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/scripts/pingcap/tidb/integrationtest_with_tikv.sh y
~/agent/workspace/pingcap/tidb/ghpr_check2/tidb/tests/integrationtest ~/agent/workspace/pingcap/tidb/ghpr_check2/tidb
extracting statistics: s
skip building tidb-server, using existing binary: /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb/bin/integration_test_tidb-server
building portgenerator binary: ./portgenerator
go: downloading go.uber.org/zap v1.27.0
go: downloading github.com/pingcap/log v1.1.1-0.20240314023424-862ccc32f18d
go: downloading github.com/phayes/freeport v0.0.0-20180830031419-95f893ade6f2
go: downloading gopkg.in/natefinch/lumberjack.v2 v2.2.1
go: downloading go.uber.org/multierr v1.11.0
[Pipeline] archiveArtifacts
Archiving artifacts
[Pipeline] archiveArtifacts
Archiving artifacts
[Pipeline] archiveArtifacts
Archiving artifacts
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest2'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest'
building mysql-tester binary: ./mysql_tester
go: downloading github.com/pingcap/mysql-tester v0.0.0-20231124064544-17b728effac3
go: downloading github.com/pingcap/errors v0.11.5-0.20221009092201-b66cddb77c32
go: downloading github.com/sirupsen/logrus v1.8.1
go: downloading github.com/defined2014/mysql v0.0.0-20231121061906-fcfacaa39f49
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest3'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pipelineddmltest'
[Pipeline] archiveArtifacts
Archiving artifacts
Sending interrupt signal to process
Killing processes
GOBIN=/home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb/tools/bin GO111MODULE=on go install github.com/pingcap/failpoint/failpoint-ctl@2eaa328
go: downloading github.com/pingcap/failpoint v0.0.0-20220801062533-2eaa32854a6c
go: downloading go.uber.org/atomic v1.11.0
go: downloading golang.org/x/sys v0.5.0
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
script returned exit code 143
script returned exit code 143
go: downloading github.com/sergi/go-diff v1.1.0
bazel --output_user_root=/home/jenkins/.tidb/tmp run --config=ci --repository_cache=/share/.cache/bazel-repository-cache //:gazelle
Extracting Bazel installation...
Starting local Bazel server and connecting to it...
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
make: *** [tools/bin/failpoint-ctl] Terminated
script returned exit code 143
make: *** [bazel_ci_simple_prepare] Terminated
kill finished with exit code 0
script returned exit code 143
script returned exit code 143
[Pipeline] // container
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
Post stage
Post stage
Post stage
Post stage
Post stage
[Pipeline] // node
[Pipeline] }
[Pipeline] }
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] junit
Recording test results
[Pipeline] junit
Recording test results
[Pipeline] junit
Recording test results
[Pipeline] junit
Recording test results
[Pipeline] junit
Recording test results
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] }
None of the test reports contained any result
[Checks API] No suitable checks publisher found.
[Pipeline] }
None of the test reports contained any result
[Checks API] No suitable checks publisher found.
None of the test reports contained any result
[Checks API] No suitable checks publisher found.
None of the test reports contained any result
[Checks API] No suitable checks publisher found.
[Pipeline] // withCredentials
None of the test reports contained any result
[Checks API] No suitable checks publisher found.
No test report files were found. Configuration error?
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] }
No test report files were found. Configuration error?
[Pipeline] // timeout
[Pipeline] // timeout
No test report files were found. Configuration error?
[Pipeline] // timeout
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
No test report files were found. Configuration error?
[Pipeline] }
No test report files were found. Configuration error?
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] }
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] // dir
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_sessiontest'
[Pipeline] // stage
[Pipeline] }
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] dir
Running in /home/jenkins/agent/workspace/pingcap/tidb/ghpr_check2/tidb
[Pipeline] {
[Pipeline] sh
[Pipeline] sh
[Pipeline] sh
+ str='run_real_tikv_tests.sh bazel_addindextest4'
+ logs_dir=logs_run_real_tikv_tests.sh_bazel_addindextest4
+ mkdir -p logs_run_real_tikv_tests.sh_bazel_addindextest4
+ mv pd1.log pd2.log pd3.log logs_run_real_tikv_tests.sh_bazel_addindextest4
+ mv tikv1.log tikv2.log tikv3.log logs_run_real_tikv_tests.sh_bazel_addindextest4
+ mv tests/integrationtest/integration-test.out logs_run_real_tikv_tests.sh_bazel_addindextest4
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_run_real_tikv_tests.sh_bazel_addindextest4.tar.gz logs_run_real_tikv_tests.sh_bazel_addindextest4
logs_run_real_tikv_tests.sh_bazel_addindextest4/
logs_run_real_tikv_tests.sh_bazel_addindextest4/tikv2.log
logs_run_real_tikv_tests.sh_bazel_addindextest4/tikv1.log
logs_run_real_tikv_tests.sh_bazel_addindextest4/pd2.log
logs_run_real_tikv_tests.sh_bazel_addindextest4/tikv3.log
logs_run_real_tikv_tests.sh_bazel_addindextest4/pd3.log
logs_run_real_tikv_tests.sh_bazel_addindextest4/pd1.log
[Pipeline] sh
+ str='integrationtest_with_tikv.sh n'
+ logs_dir=logs_integrationtest_with_tikv.sh_n
+ mkdir -p logs_integrationtest_with_tikv.sh_n
+ mv pd1.log pd2.log pd3.log logs_integrationtest_with_tikv.sh_n
+ mv tikv1.log tikv2.log tikv3.log logs_integrationtest_with_tikv.sh_n
+ mv tests/integrationtest/integration-test.out logs_integrationtest_with_tikv.sh_n
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_integrationtest_with_tikv.sh_n.tar.gz logs_integrationtest_with_tikv.sh_n
logs_integrationtest_with_tikv.sh_n/
logs_integrationtest_with_tikv.sh_n/tikv3.log
logs_integrationtest_with_tikv.sh_n/pd1.log
logs_integrationtest_with_tikv.sh_n/tikv2.log
logs_integrationtest_with_tikv.sh_n/tikv1.log
logs_integrationtest_with_tikv.sh_n/pd2.log
logs_integrationtest_with_tikv.sh_n/pd3.log
[Pipeline] sh
+ str='run_real_tikv_tests.sh bazel_flashbacktest'
+ logs_dir=logs_run_real_tikv_tests.sh_bazel_flashbacktest
+ mkdir -p logs_run_real_tikv_tests.sh_bazel_flashbacktest
+ mv pd1.log pd2.log pd3.log logs_run_real_tikv_tests.sh_bazel_flashbacktest
+ mv tikv1.log tikv2.log tikv3.log logs_run_real_tikv_tests.sh_bazel_flashbacktest
+ mv tests/integrationtest/integration-test.out logs_run_real_tikv_tests.sh_bazel_flashbacktest
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_run_real_tikv_tests.sh_bazel_flashbacktest.tar.gz logs_run_real_tikv_tests.sh_bazel_flashbacktest
logs_run_real_tikv_tests.sh_bazel_flashbacktest/
logs_run_real_tikv_tests.sh_bazel_flashbacktest/pd2.log
logs_run_real_tikv_tests.sh_bazel_flashbacktest/pd3.log
logs_run_real_tikv_tests.sh_bazel_flashbacktest/tikv1.log
logs_run_real_tikv_tests.sh_bazel_flashbacktest/pd1.log
logs_run_real_tikv_tests.sh_bazel_flashbacktest/tikv2.log
logs_run_real_tikv_tests.sh_bazel_flashbacktest/tikv3.log
[Pipeline] // container
[Pipeline] }
+ str='run_real_tikv_tests.sh bazel_importintotest2'
+ logs_dir=logs_run_real_tikv_tests.sh_bazel_importintotest2
+ mkdir -p logs_run_real_tikv_tests.sh_bazel_importintotest2
+ mv pd1.log pd2.log pd3.log logs_run_real_tikv_tests.sh_bazel_importintotest2
+ mv tikv1.log tikv2.log tikv3.log logs_run_real_tikv_tests.sh_bazel_importintotest2
+ mv tests/integrationtest/integration-test.out logs_run_real_tikv_tests.sh_bazel_importintotest2
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_run_real_tikv_tests.sh_bazel_importintotest2.tar.gz logs_run_real_tikv_tests.sh_bazel_importintotest2
logs_run_real_tikv_tests.sh_bazel_importintotest2/
logs_run_real_tikv_tests.sh_bazel_importintotest2/tikv3.log
logs_run_real_tikv_tests.sh_bazel_importintotest2/pd1.log
logs_run_real_tikv_tests.sh_bazel_importintotest2/tikv2.log
logs_run_real_tikv_tests.sh_bazel_importintotest2/pd3.log
logs_run_real_tikv_tests.sh_bazel_importintotest2/pd2.log
logs_run_real_tikv_tests.sh_bazel_importintotest2/tikv1.log
[Pipeline] // container
[Pipeline] // container
[Pipeline] }
[Pipeline] }
+ str='integrationtest_with_tikv.sh y'
+ logs_dir=logs_integrationtest_with_tikv.sh_y
+ mkdir -p logs_integrationtest_with_tikv.sh_y
+ mv pd1.log pd2.log pd3.log logs_integrationtest_with_tikv.sh_y
+ mv tikv1.log tikv2.log tikv3.log logs_integrationtest_with_tikv.sh_y
+ mv tests/integrationtest/integration-test.out logs_integrationtest_with_tikv.sh_y
mv: cannot stat 'tests/integrationtest/integration-test.out': No such file or directory
+ true
+ tar -czvf logs_integrationtest_with_tikv.sh_y.tar.gz logs_integrationtest_with_tikv.sh_y
logs_integrationtest_with_tikv.sh_y/
logs_integrationtest_with_tikv.sh_y/tikv1.log
logs_integrationtest_with_tikv.sh_y/pd2.log
logs_integrationtest_with_tikv.sh_y/pd1.log
logs_integrationtest_with_tikv.sh_y/tikv3.log
logs_integrationtest_with_tikv.sh_y/tikv2.log
logs_integrationtest_with_tikv.sh_y/pd3.log
[Pipeline] // container
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] archiveArtifacts
Archiving artifacts
[Pipeline] archiveArtifacts
Archiving artifacts
[Pipeline] archiveArtifacts
Archiving artifacts
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] archiveArtifacts
Archiving artifacts
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] }
[Pipeline] archiveArtifacts
Archiving artifacts
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] }
[Pipeline] // node
[Pipeline] // node
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // podTemplate
[Pipeline] // dir
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // dir
[Pipeline] }
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // withCredentials
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_brietest'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_pessimistictest'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest3'
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_txntest'
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] }
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_statisticstest'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest'
[Pipeline] // stage
[Pipeline] // timeout
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest1'
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // container
[Pipeline] // container
[Pipeline] }
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // node
[Pipeline] // node
[Pipeline] }
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh n'
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_addindextest4'
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_flashbacktest'
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest2'
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - SCRIPT_AND_ARGS = 'integrationtest_with_tikv.sh y'
[Pipeline] // parallel
[Pipeline] }
[Pipeline] // stage
[Pipeline] stage
[Pipeline] { (Declarative: Post Actions)
[Pipeline] container
[Pipeline] {
[Pipeline] sh
+ bash scripts/plugins/report_job_result.sh ABORTED result.json
http://fileserver.pingcap.net
--2024-05-06 03:46:01--  http://fileserver.pingcap.net/download/rd-atom-agent/agent_upload_verifyci_metadata.py
Resolving fileserver.pingcap.net (fileserver.pingcap.net)... 10.2.12.82
Connecting to fileserver.pingcap.net (fileserver.pingcap.net)|10.2.12.82|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 4181 (4.1K) [application/octet-stream]
Saving to: ‘agent_upload_verifyci_metadata.py’

     0K ....                                                  100%  630M=0s

2024-05-06 03:46:02 (630 MB/s) - ‘agent_upload_verifyci_metadata.py’ saved [4181/4181]

No junit report file
parse result file result.json success
upload data succesfully.
[Pipeline] }
[Pipeline] // container
[Pipeline] archiveArtifacts
Archiving artifacts
Recording fingerprints
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] End of Pipeline
org.jenkinsci.plugins.workflow.actions.ErrorAction$ErrorId: 65ba8b7c-6e25-4fcf-b128-a3cce1917849
Also:   org.jenkinsci.plugins.workflow.actions.ErrorAction$ErrorId: aef33a9c-9ce3-4c49-b0be-69a5f060fa02
org.jenkinsci.plugins.workflow.support.steps.AgentOfflineException: Unable to create live FilePath for pingcap-tidb-ghpr-check2-8824-xhw47-8ndpj-58jtj
	at org.jenkinsci.plugins.workflow.support.steps.ExecutorStepDynamicContext$FilePathTranslator.get(ExecutorStepDynamicContext.java:184)
	at org.jenkinsci.plugins.workflow.support.steps.ExecutorStepDynamicContext$FilePathTranslator.get(ExecutorStepDynamicContext.java:156)
	at org.jenkinsci.plugins.workflow.support.steps.ExecutorStepDynamicContext$Translator.get(ExecutorStepDynamicContext.java:149)
	at org.jenkinsci.plugins.workflow.support.steps.ExecutorStepDynamicContext$FilePathTranslator.get(ExecutorStepDynamicContext.java:166)
	at org.jenkinsci.plugins.workflow.support.steps.ExecutorStepDynamicContext$FilePathTranslator.get(ExecutorStepDynamicContext.java:156)
	at org.jenkinsci.plugins.workflow.steps.DynamicContext$Typed.get(DynamicContext.java:95)
	at org.jenkinsci.plugins.workflow.cps.ContextVariableSet.get(ContextVariableSet.java:139)
	at org.jenkinsci.plugins.workflow.cps.CpsThread.getContextVariable(CpsThread.java:137)
	at org.jenkinsci.plugins.workflow.cps.CpsStepContext.doGet(CpsStepContext.java:297)
	at org.jenkinsci.plugins.workflow.support.DefaultStepContext.get(DefaultStepContext.java:75)
	at org.jenkinsci.plugins.workflow.steps.StepDescriptor.checkContextAvailability(StepDescriptor.java:263)
	at org.jenkinsci.plugins.workflow.cps.DSL.invokeStep(DSL.java:300)
	at org.jenkinsci.plugins.workflow.cps.DSL.invokeMethod(DSL.java:196)
	at org.jenkinsci.plugins.workflow.cps.CpsScript.invokeMethod(CpsScript.java:124)
	at org.codehaus.groovy.runtime.callsite.PogoMetaClassSite.call(PogoMetaClassSite.java:47)
	at org.codehaus.groovy.runtime.callsite.CallSiteArray.defaultCall(CallSiteArray.java:47)
	at org.codehaus.groovy.runtime.callsite.AbstractCallSite.call(AbstractCallSite.java:116)
	at com.cloudbees.groovy.cps.sandbox.DefaultInvoker.methodCall(DefaultInvoker.java:20)
	at org.jenkinsci.plugins.workflow.cps.LoggingInvoker.methodCall(LoggingInvoker.java:105)
	at org.jenkinsci.plugins.pipeline.modeldefinition.agent.CheckoutScript.performCheckout(CheckoutScript.groovy:77)
	at org.jenkinsci.plugins.pipeline.modeldefinition.agent.CheckoutScript.checkoutAndRun(CheckoutScript.groovy:56)
	at org.jenkinsci.plugins.pipeline.modeldefinition.agent.CheckoutScript.doCheckout(CheckoutScript.groovy:40)
	at org.csanchez.jenkins.plugins.kubernetes.pipeline.KubernetesDeclarativeAgentScript.run(KubernetesDeclarativeAgentScript.groovy:54)
	at ___cps.transform___(Native Method)
	at com.cloudbees.groovy.cps.impl.ContinuationGroup.methodCall(ContinuationGroup.java:90)
	at com.cloudbees.groovy.cps.impl.FunctionCallBlock$ContinuationImpl.dispatchOrArg(FunctionCallBlock.java:116)
	at com.cloudbees.groovy.cps.impl.FunctionCallBlock$ContinuationImpl.fixArg(FunctionCallBlock.java:85)
	at jdk.internal.reflect.GeneratedMethodAccessor209.invoke(Unknown Source)
	at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.base/java.lang.reflect.Method.invoke(Method.java:566)
	at com.cloudbees.groovy.cps.impl.ContinuationPtr$ContinuationImpl.receive(ContinuationPtr.java:72)
	at com.cloudbees.groovy.cps.impl.PropertyishBlock$ContinuationImpl.get(PropertyishBlock.java:75)
	at com.cloudbees.groovy.cps.LValueBlock$GetAdapter.receive(LValueBlock.java:30)
	at com.cloudbees.groovy.cps.impl.PropertyishBlock$ContinuationImpl.fixName(PropertyishBlock.java:65)
	at jdk.internal.reflect.GeneratedMethodAccessor760.invoke(Unknown Source)
	at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.base/java.lang.reflect.Method.invoke(Method.java:566)
	at com.cloudbees.groovy.cps.impl.ContinuationPtr$ContinuationImpl.receive(ContinuationPtr.java:72)
	at com.cloudbees.groovy.cps.impl.ConstantBlock.eval(ConstantBlock.java:21)
	at com.cloudbees.groovy.cps.Next.step(Next.java:83)
	at com.cloudbees.groovy.cps.Continuable$1.call(Continuable.java:152)
	at com.cloudbees.groovy.cps.Continuable$1.call(Continuable.java:146)
	at org.codehaus.groovy.runtime.GroovyCategorySupport$ThreadCategoryInfo.use(GroovyCategorySupport.java:136)
	at org.codehaus.groovy.runtime.GroovyCategorySupport.use(GroovyCategorySupport.java:275)
	at com.cloudbees.groovy.cps.Continuable.run0(Continuable.java:146)
	at org.jenkinsci.plugins.workflow.cps.SandboxContinuable.access$001(SandboxContinuable.java:18)
	at org.jenkinsci.plugins.workflow.cps.SandboxContinuable.run0(SandboxContinuable.java:51)
	at org.jenkinsci.plugins.workflow.cps.CpsThread.runNextChunk(CpsThread.java:187)
	at org.jenkinsci.plugins.workflow.cps.CpsThreadGroup.run(CpsThreadGroup.java:423)
	at org.jenkinsci.plugins.workflow.cps.CpsThreadGroup$2.call(CpsThreadGroup.java:331)
	at org.jenkinsci.plugins.workflow.cps.CpsThreadGroup$2.call(CpsThreadGroup.java:295)
	at org.jenkinsci.plugins.workflow.cps.CpsVmExecutorService$2.call(CpsVmExecutorService.java:97)
	at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
	at hudson.remoting.SingleLaneExecutorService$1.run(SingleLaneExecutorService.java:139)
	at jenkins.util.ContextResettingExecutorService$1.run(ContextResettingExecutorService.java:28)
	at jenkins.security.ImpersonatingExecutorService$1.run(ImpersonatingExecutorService.java:68)
	at jenkins.util.ErrorLoggingExecutorService.lambda$wrap$0(ErrorLoggingExecutorService.java:51)
	at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
	at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
	at java.base/java.lang.Thread.run(Thread.java:829)
script returned exit code 2
Failed in branch Matrix - SCRIPT_AND_ARGS = 'run_real_tikv_tests.sh bazel_importintotest4'
org.jenkinsci.plugins.workflow.actions.ErrorAction$ErrorId: 9621c347-dbfd-401d-9a52-bdcb21b6c08e
Finished: ABORTED