Skip to content

Console Output

Skipping 2,156 KB.. Full Log
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
table consistent_replicate_storage_file.check1 not exists for 4-th check, retry later
check_processor_table_count http://127.0.0.1:2379 986bbba8-7dd5-4a45-b561-f10803268091 573cfc7e-7e9e-4783-9bca-3e8934d99b2f 0
table ddl_manager.finish_mark not exists for 75-th check, retry later
run task successfully
+ set +x
+ tso='449617742475296769
PASS
coverage: 2.0% of statements in github.com/pingcap/tiflow/...'
+ echo 449617742475296769 PASS coverage: 2.0% of statements in github.com/pingcap/tiflow/...
+ awk -F ' ' '{print $1}'
+ set +x
[Wed May  8 16:06:14 CST 2024] <<<<<< START cdc server in multi_cdc_cluster case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ '[' -z '' ']'
+ GO_FAILPOINTS=
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL'
+ [[ no != \n\o ]]
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.multi_cdc_cluster.1314613148.out server --log-file /tmp/tidb_cdc_test/multi_cdc_cluster/cdcmult_cdc.server1.log --log-level debug --data-dir /tmp/tidb_cdc_test/multi_cdc_cluster/cdc_datamult_cdc.server1 --cluster-id test1 --addr 127.0.0.1:8300
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8300; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
{"level":"warn","ts":"2024-05-08T16:06:15.113572+0800","logger":"etcd-client","caller":"v3@v3.5.12/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc000afb180/127.0.0.1:2479","attempt":0,"error":"rpc error: code = DeadlineExceeded desc = context deadline exceeded"}
table owner_remove_table_error.finished_mark not exists for 1-th check, retry later
***************** properties *****************
"insertproportion"="0"
"dotransactions"="false"
"workload"="core"
"requestdistribution"="uniform"
"updateproportion"="0"
"mysql.db"="changefeed_reconstruct"
"mysql.port"="4000"
"readallfields"="true"
"threadcount"="4"
"recordcount"="50"
"operationcount"="0"
"mysql.host"="127.0.0.1"
"mysql.user"="root"
"scanproportion"="0"
"readproportion"="0"
**********************************************
Run finished, takes 24.293494ms
INSERT - Takes(s): 0.0, Count: 48, OPS: 2521.9, Avg(us): 1878, Min(us): 1215, Max(us): 5284, 95th(us): 6000, 99th(us): 6000
table changefeed_reconstruct.usertable not exists for 1-th check, retry later
table consistent_replicate_storage_file.check1 not exists for 5-th check, retry later
check diff failed 4-th time, retry later
table ddl_manager.finish_mark not exists for 76-th check, retry later
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Wed, 08 May 2024 08:06:16 GMT
< Content-Length: 827
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/31c4124a-4030-47cf-83dc-92798d22dda7
	{"id":"31c4124a-4030-47cf-83dc-92798d22dda7","address":"127.0.0.1:8300","version":"v8.1.0-alpha-29-gee986f33f","git-hash":"ee986f33f2b2c4927b5c2c0de0e6242a0e9f7db3","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/bin/cdc.test","start-timestamp":1715155573}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f573cd6fc4c
	31c4124a-4030-47cf-83dc-92798d22dda7

/tidb/cdc/default/default/upstream/7366537032084384495
	{"id":7366537032084384495,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/31c4124a-4030-47cf-83dc-92798d22dda7
	{"id":"31c4124a-4030-47cf-83dc-92798d22dda7","address":"127.0.0.1:8300","version":"v8.1.0-alpha-29-gee986f33f","git-hash":"ee986f33f2b2c4927b5c2c0de0e6242a0e9f7db3","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/bin/cdc.test","start-timestamp":1715155573}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f573cd6fc4c
	31c4124a-4030-47cf-83dc-92798d22dda7

/tidb/cdc/default/default/upstream/7366537032084384495
	{"id":7366537032084384495,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/default/__cdc_meta__/capture/31c4124a-4030-47cf-83dc-92798d22dda7
	{"id":"31c4124a-4030-47cf-83dc-92798d22dda7","address":"127.0.0.1:8300","version":"v8.1.0-alpha-29-gee986f33f","git-hash":"ee986f33f2b2c4927b5c2c0de0e6242a0e9f7db3","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/bin/cdc.test","start-timestamp":1715155573}

/tidb/cdc/default/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/default/__cdc_meta__/owner/22318f573cd6fc4c
	31c4124a-4030-47cf-83dc-92798d22dda7

/tidb/cdc/default/default/upstream/7366537032084384495
	{"id":7366537032084384495,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.ddl_puller_lag.cli.5171.out cli changefeed create --start-ts=449617742239629313 --sink-uri=mysql+ssl://normal:123456@127.0.0.1:3306/
[2024/05/08 16:06:16.675 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc0018d43c0/127.0.0.1:2379] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
[2024/05/08 16:06:16.687 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc000c343c0/127.0.0.1:2479] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
table owner_remove_table_error.finished_mark not exists for 2-th check, retry later
Create changefeed successfully!
ID: 3bf0a74c-98e9-4f79-abac-0485bfc52036
Info: {"upstream_id":7366537032084384495,"namespace":"default","id":"3bf0a74c-98e9-4f79-abac-0485bfc52036","sink_uri":"mysql+ssl://normal:xxxxx@127.0.0.1:3306/","create_time":"2024-05-08T16:06:17.110215363+08:00","start_ts":449617742239629313,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["*.*"]},"mounter":{"worker_num":16},"sink":{"csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.1.0-alpha-29-gee986f33f","resolved_ts":449617742239629313,"checkpoint_ts":449617742239629313,"checkpoint_time":"2024-05-08 16:06:11.898"}
PASS
coverage: 2.6% of statements in github.com/pingcap/tiflow/...
[2024/05/08 16:06:16.681 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc0016203c0/127.0.0.1:2479] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
[2024/05/08 16:06:16.685 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc001ec83c0/127.0.0.1:2379] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
table changefeed_reconstruct.usertable exists
check diff failed 1-th time, retry later
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8300/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8300 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8300 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8300
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Wed, 08 May 2024 08:06:17 GMT
< Content-Length: 819
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/test1/__cdc_meta__/capture/739116fd-515d-4975-bc9a-4bbb9fcc679f
	{"id":"739116fd-515d-4975-bc9a-4bbb9fcc679f","address":"127.0.0.1:8300","version":"v8.1.0-alpha-29-gee986f33f","git-hash":"ee986f33f2b2c4927b5c2c0de0e6242a0e9f7db3","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/bin/cdc.test","start-timestamp":1715155574}

/tidb/cdc/test1/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/test1/__cdc_meta__/owner/22318f573cdc734f
	739116fd-515d-4975-bc9a-4bbb9fcc679f

/tidb/cdc/test1/default/upstream/7366537030980399069
	{"id":7366537030980399069,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/test1/__cdc_meta__/capture/739116fd-515d-4975-bc9a-4bbb9fcc679f
	{"id":"739116fd-515d-4975-bc9a-4bbb9fcc679f","address":"127.0.0.1:8300","version":"v8.1.0-alpha-29-gee986f33f","git-hash":"ee986f33f2b2c4927b5c2c0de0e6242a0e9f7db3","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/bin/cdc.test","start-timestamp":1715155574}

/tidb/cdc/test1/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/test1/__cdc_meta__/owner/22318f573cdc734f
	739116fd-515d-4975-bc9a-4bbb9fcc679f

/tidb/cdc/test1/default/upstream/7366537030980399069
	{"id":7366537030980399069,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/test1/__cdc_meta__/capture/739116fd-515d-4975-bc9a-4bbb9fcc679f
	{"id":"739116fd-515d-4975-bc9a-4bbb9fcc679f","address":"127.0.0.1:8300","version":"v8.1.0-alpha-29-gee986f33f","git-hash":"ee986f33f2b2c4927b5c2c0de0e6242a0e9f7db3","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/bin/cdc.test","start-timestamp":1715155574}

/tidb/cdc/test1/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/test1/__cdc_meta__/owner/22318f573cdc734f
	739116fd-515d-4975-bc9a-4bbb9fcc679f

/tidb/cdc/test1/default/upstream/7366537030980399069
	{"id":7366537030980399069,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
[Wed May  8 16:06:17 CST 2024] <<<<<< START cdc server in multi_cdc_cluster case >>>>>>
+ [[ '' == \t\r\u\e ]]
+ set +e
+ get_info_fail_msg='failed to get info:'
+ etcd_info_msg='etcd info'
+ GO_FAILPOINTS=
+ '[' -z '' ']'
+ curl_status_cmd='curl -vsL --max-time 20 http://127.0.0.1:8301/debug/info --user ticdc:ticdc_secret -vsL'
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.multi_cdc_cluster.1319813200.out server --log-file /tmp/tidb_cdc_test/multi_cdc_cluster/cdcmult_cdc.server2.log --log-level debug --data-dir /tmp/tidb_cdc_test/multi_cdc_cluster/cdc_datamult_cdc.server2 --cluster-id test2 --addr 127.0.0.1:8301
+ [[ no != \n\o ]]
+ (( i = 0 ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8301/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8301 (#0)
*   Trying 127.0.0.1...
* Connection refused
* Failed connect to 127.0.0.1:8301; Connection refused
* Closing connection 0
+ res=
+ echo ''
+ grep -q 'failed to get info:'
+ echo ''
+ grep -q 'etcd info'
+ '[' 0 -eq 50 ']'
+ sleep 3
table consistent_replicate_storage_file.check1 not exists for 6-th check, retry later
check diff failed 5-th time, retry later
+ set +x
table sink_retry.finish_mark_2 exists
check diff successfully
wait process cdc.test exit for 1-th time...
wait process cdc.test exit for 2-th time...
wait process cdc.test exit for 3-th time...
cdc.test: no process found
wait process cdc.test exit for 4-th time...
process cdc.test already exit
[Wed May  8 16:06:17 CST 2024] <<<<<< run test case sink_retry success! >>>>>>
table ddl_manager.finish_mark not exists for 77-th check, retry later
table owner_remove_table_error.finished_mark exists
check diff successfully
[2024/05/08 16:06:19.178 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc001ec83c0/127.0.0.1:2379] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
[2024/05/08 16:06:19.028 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc0018d43c0/127.0.0.1:2379] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
[2024/05/08 16:06:19.171 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc000c343c0/127.0.0.1:2479] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
wait process cdc.test exit for 1-th time...
table consistent_replicate_storage_file.check1 not exists for 7-th check, retry later
check diff successfully
table ddl_manager.finish_mark not exists for 78-th check, retry later
wait process cdc.test exit for 2-th time...
check diff successfully
wait process cdc.test exit for 3-th time...
wait process cdc.test exit for 1-th time...
[2024/05/08 16:06:20.677 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc0016d6000/127.0.0.1:2379] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
[2024/05/08 16:06:20.820 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc0016d6000/127.0.0.1:2379] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
+ (( i++ ))
+ (( i <= 50 ))
++ curl -vsL --max-time 20 http://127.0.0.1:8301/debug/info --user ticdc:ticdc_secret -vsL
* About to connect() to 127.0.0.1 port 8301 (#0)
*   Trying 127.0.0.1...
* Connected to 127.0.0.1 (127.0.0.1) port 8301 (#0)
* Server auth using Basic with user 'ticdc'
> GET /debug/info HTTP/1.1
> Authorization: Basic dGljZGM6dGljZGNfc2VjcmV0
> User-Agent: curl/7.29.0
> Host: 127.0.0.1:8301
> Accept: */*
> 
< HTTP/1.1 200 OK
< Date: Wed, 08 May 2024 08:06:20 GMT
< Content-Length: 819
< Content-Type: text/plain; charset=utf-8
< 
{ [data not shown]
* Connection #0 to host 127.0.0.1 left intact
+ res='

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/test2/__cdc_meta__/capture/d25c8518-4686-46a1-9f0d-3e528b8c943b
	{"id":"d25c8518-4686-46a1-9f0d-3e528b8c943b","address":"127.0.0.1:8301","version":"v8.1.0-alpha-29-gee986f33f","git-hash":"ee986f33f2b2c4927b5c2c0de0e6242a0e9f7db3","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/bin/cdc.test","start-timestamp":1715155577}

/tidb/cdc/test2/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/test2/__cdc_meta__/owner/22318f573cdc7375
	d25c8518-4686-46a1-9f0d-3e528b8c943b

/tidb/cdc/test2/default/upstream/7366537030980399069
	{"id":7366537030980399069,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/test2/__cdc_meta__/capture/d25c8518-4686-46a1-9f0d-3e528b8c943b
	{"id":"d25c8518-4686-46a1-9f0d-3e528b8c943b","address":"127.0.0.1:8301","version":"v8.1.0-alpha-29-gee986f33f","git-hash":"ee986f33f2b2c4927b5c2c0de0e6242a0e9f7db3","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/bin/cdc.test","start-timestamp":1715155577}

/tidb/cdc/test2/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/test2/__cdc_meta__/owner/22318f573cdc7375
	d25c8518-4686-46a1-9f0d-3e528b8c943b

/tidb/cdc/test2/default/upstream/7366537030980399069
	{"id":7366537030980399069,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'failed to get info:'
+ echo '

*** owner info ***:



*** processors info ***:



*** etcd info ***:

/tidb/cdc/test2/__cdc_meta__/capture/d25c8518-4686-46a1-9f0d-3e528b8c943b
	{"id":"d25c8518-4686-46a1-9f0d-3e528b8c943b","address":"127.0.0.1:8301","version":"v8.1.0-alpha-29-gee986f33f","git-hash":"ee986f33f2b2c4927b5c2c0de0e6242a0e9f7db3","deploy-path":"/home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/bin/cdc.test","start-timestamp":1715155577}

/tidb/cdc/test2/__cdc_meta__/meta/meta-version
	1

/tidb/cdc/test2/__cdc_meta__/owner/22318f573cdc7375
	d25c8518-4686-46a1-9f0d-3e528b8c943b

/tidb/cdc/test2/default/upstream/7366537030980399069
	{"id":7366537030980399069,"pd-endpoints":"http://127.0.0.1:2379,http://127.0.0.1:2379","key-path":"","cert-path":"","ca-path":"","cert-allowed-cn":null}'
+ grep -q 'etcd info'
+ break
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.multi_cdc_cluster.cli.13253.out cli changefeed create --start-ts=449617742475296769 --sink-uri=mysql://normal:123456@127.0.0.1:3306/ --server http://127.0.0.1:8300 --config=/home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/tests/integration_tests/multi_cdc_cluster/conf/changefeed1.toml
cdc.test: no process found
wait process cdc.test exit for 4-th time...
process cdc.test already exit
[Wed May  8 16:06:21 CST 2024] <<<<<< run test case owner_remove_table_error success! >>>>>>
wait process cdc.test exit for 2-th time...
Create changefeed successfully!
ID: aa929cf2-4130-4238-b3c4-2ddb5e758b4b
Info: {"upstream_id":7366537030980399069,"namespace":"default","id":"aa929cf2-4130-4238-b3c4-2ddb5e758b4b","sink_uri":"mysql://normal:xxxxx@127.0.0.1:3306/","create_time":"2024-05-08T16:06:21.157557946+08:00","start_ts":449617742475296769,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["test.multi_cdc1"]},"mounter":{"worker_num":16},"sink":{"csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.1.0-alpha-29-gee986f33f","resolved_ts":449617742475296769,"checkpoint_ts":449617742475296769,"checkpoint_time":"2024-05-08 16:06:12.797"}
PASS
coverage: 2.6% of statements in github.com/pingcap/tiflow/...
wait process cdc.test exit for 3-th time...
[2024/05/08 16:06:21.468 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc0018d43c0/127.0.0.1:2379] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
[2024/05/08 16:06:21.679 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc000c343c0/127.0.0.1:2479] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
check diff failed 1-th time, retry later
cdc.test: no process found
wait process cdc.test exit for 4-th time...
process cdc.test already exit
[Wed May  8 16:06:21 CST 2024] <<<<<< run test case changefeed_reconstruct success! >>>>>>
table consistent_replicate_storage_file.check1 not exists for 8-th check, retry later
[2024/05/08 16:06:21.968 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc001ec83c0/127.0.0.1:2379] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
[2024/05/08 16:06:22.290 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc001bb8000/127.0.0.1:2379] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
[2024/05/08 16:06:22.290 +08:00] [WARN] [retry_interceptor.go:62] ["retrying of unary invoker failed"] [target=etcd-endpoints://0xc0017383c0/127.0.0.1:2379] [attempt=0] [error="rpc error: code = DeadlineExceeded desc = context deadline exceeded"]
table ddl_manager.finish_mark not exists for 79-th check, retry later
+ set +x
+ cdc.test -test.coverprofile=/tmp/tidb_cdc_test/cov.multi_cdc_cluster.cli.13289.out cli changefeed create --start-ts=449617742475296769 --sink-uri=mysql://normal:123456@127.0.0.1:3306/ --server http://127.0.0.1:8301 --config=/home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/tests/integration_tests/multi_cdc_cluster/conf/changefeed2.toml
Create changefeed successfully!
ID: bcd4d85a-556c-466d-bf33-0653f7b8446f
Info: {"upstream_id":7366537030980399069,"namespace":"default","id":"bcd4d85a-556c-466d-bf33-0653f7b8446f","sink_uri":"mysql://normal:xxxxx@127.0.0.1:3306/","create_time":"2024-05-08T16:06:23.079313807+08:00","start_ts":449617742475296769,"config":{"memory_quota":1073741824,"case_sensitive":false,"force_replicate":false,"ignore_ineligible_table":false,"check_gc_safe_point":true,"enable_sync_point":false,"enable_table_monitor":false,"bdr_mode":false,"sync_point_interval":600000000000,"sync_point_retention":86400000000000,"filter":{"rules":["test.multi_cdc2"]},"mounter":{"worker_num":16},"sink":{"csv":{"delimiter":",","quote":"\"","null":"\\N","include_commit_ts":false,"binary_encoding_method":"base64","output_old_value":false,"output_handle_key":false},"encoder_concurrency":32,"terminator":"\r\n","date_separator":"day","enable_partition_separator":true,"enable_kafka_sink_v2":false,"only_output_updated_columns":false,"delete_only_output_handle_key_columns":false,"content_compatible":false,"advance_timeout":150,"send_bootstrap_interval_in_sec":120,"send_bootstrap_in_msg_count":10000,"send_bootstrap_to_all_partition":true,"debezium_disable_schema":false,"debezium":{"output_old_value":true},"open":{"output_old_value":true}},"consistent":{"level":"none","max_log_size":64,"flush_interval":2000,"meta_flush_interval":200,"encoding_worker_num":16,"flush_worker_num":8,"use_file_backend":false,"memory_usage":{"memory_quota_percentage":50}},"scheduler":{"enable_table_across_nodes":false,"region_threshold":100000,"write_key_threshold":0},"integrity":{"integrity_check_level":"none","corruption_handle_level":"warn"},"changefeed_error_stuck_duration":1800000000000,"synced_status":{"synced_check_interval":300,"checkpoint_interval":15}},"state":"normal","creator_version":"v8.1.0-alpha-29-gee986f33f","resolved_ts":449617742475296769,"checkpoint_ts":449617742475296769,"checkpoint_time":"2024-05-08 16:06:12.797"}
PASS
coverage: 2.6% of statements in github.com/pingcap/tiflow/...
check diff failed 2-th time, retry later
table consistent_replicate_storage_file.check1 not exists for 9-th check, retry later
table ddl_manager.finish_mark not exists for 80-th check, retry later
+ set +x
table test.multi_cdc1 exists
table test.multi_cdc2 exists
check diff failed 1-th time, retry later
[2024/05/08 16:06:16.653 +08:00] [WARN] [diff.go:182] ["table struct is not equal"] [reason="column num not equal, one is 34 another is 20"]
[2024/05/08 16:06:21.655 +08:00] [WARN] [diff.go:182] ["table struct is not equal"] [reason="column num not equal, one is 34 another is 28"]
table changefeed_dup_error_restart.finish_mark_2 exists
check diff successfully
wait process cdc.test exit for 1-th time...
wait process cdc.test exit for 2-th time...
wait process cdc.test exit for 3-th time...
cdc.test: no process found
wait process cdc.test exit for 4-th time...
process cdc.test already exit
[Wed May  8 16:06:24 CST 2024] <<<<<< run test case changefeed_dup_error_restart success! >>>>>>
table consistent_replicate_storage_file.check1 not exists for 10-th check, retry later
check diff failed 3-th time, retry later
table ddl_manager.finish_mark not exists for 81-th check, retry later
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/tests/integration_tests/changefeed_error/run.sh using Sink-Type: mysql... <<=================
The 1 times to try to start tidb cluster...
check diff successfully
wait process cdc.test exit for 1-th time...
table consistent_replicate_storage_file.check1 not exists for 11-th check, retry later
check diff failed 4-th time, retry later
wait process cdc.test exit for 2-th time...
table ddl_manager.finish_mark not exists for 82-th check, retry later
wait process cdc.test exit for 3-th time...
cdc.test: no process found
wait process cdc.test exit for 4-th time...
process cdc.test already exit
[Wed May  8 16:06:28 CST 2024] <<<<<< run test case multi_cdc_cluster success! >>>>>>
start tidb cluster in /tmp/tidb_cdc_test/changefeed_error
Starting Upstream PD...
Release Version: v8.1.0-alpha-20-g13eaf8aaa
Edition: Community
Git Commit Hash: 13eaf8aaa661dc0d3885ac9e59145f69752cee2d
Git Branch: release-8.1
UTC Build Time:  2024-04-30 07:22:55
check diff failed 5-th time, retry later
table ddl_manager.finish_mark not exists for 83-th check, retry later
Starting Downstream PD...
Release Version: v8.1.0-alpha-20-g13eaf8aaa
Edition: Community
Git Commit Hash: 13eaf8aaa661dc0d3885ac9e59145f69752cee2d
Git Branch: release-8.1
UTC Build Time:  2024-04-30 07:22:55
Verifying upstream PD is started...
table consistent_replicate_storage_file.check1 not exists for 12-th check, retry later
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/tests/integration_tests/multi_capture/run.sh using Sink-Type: mysql... <<=================
The 1 times to try to start tidb cluster...
check diff failed 6-th time, retry later
table consistent_replicate_storage_file.check1 not exists for 13-th check, retry later
table ddl_manager.finish_mark not exists for 84-th check, retry later
table consistent_replicate_storage_file.check1 not exists for 14-th check, retry later
Verifying downstream PD is started...
Starting Upstream TiKV...
TiKV 
Release Version:   8.1.0
Edition:           Community
Git Commit Hash:   56613f7c3e28c02853cc51d15bc1b77f68b58be8
Git Commit Branch: release-8.1
UTC Build Time:    2024-04-30 06:15:40
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.1.0
Edition:           Community
Git Commit Hash:   56613f7c3e28c02853cc51d15bc1b77f68b58be8
Git Commit Branch: release-8.1
UTC Build Time:    2024-04-30 06:15:40
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
check diff successfully
table ddl_manager.finish_mark not exists for 85-th check, retry later
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] // cache
[Pipeline] }
[Pipeline] // dir
Post stage
[Pipeline] sh
start tidb cluster in /tmp/tidb_cdc_test/multi_capture
Starting Upstream PD...
Release Version: v8.1.0-alpha-20-g13eaf8aaa
Edition: Community
Git Commit Hash: 13eaf8aaa661dc0d3885ac9e59145f69752cee2d
Git Branch: release-8.1
UTC Build Time:  2024-04-30 07:22:55
Starting Downstream PD...
Release Version: v8.1.0-alpha-20-g13eaf8aaa
Edition: Community
Git Commit Hash: 13eaf8aaa661dc0d3885ac9e59145f69752cee2d
Git Branch: release-8.1
UTC Build Time:  2024-04-30 07:22:55
Verifying upstream PD is started...
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/tests/integration_tests/kafka_big_messages/run.sh using Sink-Type: mysql... <<=================
[Wed May  8 16:06:34 CST 2024] <<<<<< run test case kafka_big_messages success! >>>>>>
Starting Upstream TiDB...
Release Version: v8.1.0-alpha-107-g238f2e688e
Edition: Community
Git Commit Hash: 238f2e688e9b69c1fa0f167de290828b7d99d172
Git Branch: release-8.1
UTC Build Time: 2024-05-07 09:31:59
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.1.0-alpha-107-g238f2e688e
Edition: Community
Git Commit Hash: 238f2e688e9b69c1fa0f167de290828b7d99d172
Git Branch: release-8.1
UTC Build Time: 2024-05-07 09:31:59
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
\033[0;36m<<< Run all test success >>>\033[0m
[Pipeline] }
Cache not saved (ws/jenkins-pingcap-tiflow-release-8.1-pull_cdc_integration_mysql_test-84/tiflow-cdc already exists)
table consistent_replicate_storage_file.check1 not exists for 15-th check, retry later
[Pipeline] // cache
[Pipeline] }
[Pipeline] // dir
Post stage
[Pipeline] sh
+ ls /tmp/tidb_cdc_test/
default_value
++ find /tmp/tidb_cdc_test/ -type f -name '*.log'
+ tar -cvzf log-G10.tar.gz /tmp/tidb_cdc_test/default_value/tikv2.log /tmp/tidb_cdc_test/default_value/tiflash/log/proxy.log /tmp/tidb_cdc_test/default_value/tiflash/log/error.log /tmp/tidb_cdc_test/default_value/tiflash/log/server.log /tmp/tidb_cdc_test/default_value/tiflash/db/proxy/db/000005.log /tmp/tidb_cdc_test/default_value/tikv_down.log /tmp/tidb_cdc_test/default_value/tikv3/db/000005.log /tmp/tidb_cdc_test/default_value/tikv1.log /tmp/tidb_cdc_test/default_value/tidb_down.log /tmp/tidb_cdc_test/default_value/tikv2/db/000005.log /tmp/tidb_cdc_test/default_value/tidb_other.log /tmp/tidb_cdc_test/default_value/tidb-slow.log /tmp/tidb_cdc_test/default_value/down_pd.log /tmp/tidb_cdc_test/default_value/tikv1/db/000005.log /tmp/tidb_cdc_test/default_value/down_pd/region-meta/000001.log /tmp/tidb_cdc_test/default_value/down_pd/hot-region/000001.log /tmp/tidb_cdc_test/default_value/pd1.log /tmp/tidb_cdc_test/default_value/tikv_down/db/000005.log /tmp/tidb_cdc_test/default_value/pd1/region-meta/000001.log /tmp/tidb_cdc_test/default_value/pd1/hot-region/000001.log /tmp/tidb_cdc_test/default_value/tidb.log /tmp/tidb_cdc_test/default_value/tikv3.log
tar: Removing leading `/' from member names
/tmp/tidb_cdc_test/default_value/tikv2.log
/tmp/tidb_cdc_test/default_value/tiflash/log/proxy.log
/tmp/tidb_cdc_test/default_value/tiflash/log/error.log
/tmp/tidb_cdc_test/default_value/tiflash/log/server.log
/tmp/tidb_cdc_test/default_value/tiflash/db/proxy/db/000005.log
/tmp/tidb_cdc_test/default_value/tikv_down.log
check diff failed 1-th time, retry later
table ddl_manager.finish_mark not exists for 86-th check, retry later
/tmp/tidb_cdc_test/default_value/tikv3/db/000005.log
/tmp/tidb_cdc_test/default_value/tikv1.log
+ ls /tmp/tidb_cdc_test/
bank
cov.bank.25842586.out
cov.bank.cli.2637.out
cov.kill_owner_with_ddl.79317933.out
cov.kill_owner_with_ddl.80728074.out
cov.kill_owner_with_ddl.82518253.out
cov.kill_owner_with_ddl.83738375.out
cov.kill_owner_with_ddl.85128514.out
cov.kill_owner_with_ddl.86788680.out
cov.owner_remove_table_error.1111711119.out
kill_owner_with_ddl
owner_remove_table_error
sql_res.bank.txt
sql_res.kill_owner_with_ddl.txt
sql_res.owner_remove_table_error.txt
++ find /tmp/tidb_cdc_test/ -type f -name '*.log'
+ tar -cvzf log-G21.tar.gz /tmp/tidb_cdc_test/bank/tikv2.log /tmp/tidb_cdc_test/bank/stdout.log /tmp/tidb_cdc_test/bank/tikv3.log /tmp/tidb_cdc_test/bank/tikv_down.log /tmp/tidb_cdc_test/bank/tikv2-2024-05-08T16-03-10.758.log /tmp/tidb_cdc_test/bank/tidb_other.log /tmp/tidb_cdc_test/bank/pd1.log /tmp/tidb_cdc_test/bank/tikv1.log /tmp/tidb_cdc_test/bank/tidb.log /tmp/tidb_cdc_test/bank/tidb-slow.log /tmp/tidb_cdc_test/bank/down_pd.log /tmp/tidb_cdc_test/bank/cdc.log /tmp/tidb_cdc_test/bank/tidb_down.log /tmp/tidb_cdc_test/kill_owner_with_ddl/tikv2.log /tmp/tidb_cdc_test/kill_owner_with_ddl/stdout.log /tmp/tidb_cdc_test/kill_owner_with_ddl/tikv3.log /tmp/tidb_cdc_test/kill_owner_with_ddl/tikv_down.log /tmp/tidb_cdc_test/kill_owner_with_ddl/tidb_other.log /tmp/tidb_cdc_test/kill_owner_with_ddl/pd1.log /tmp/tidb_cdc_test/kill_owner_with_ddl/tikv1.log /tmp/tidb_cdc_test/kill_owner_with_ddl/tidb.log /tmp/tidb_cdc_test/kill_owner_with_ddl/tidb-slow.log /tmp/tidb_cdc_test/kill_owner_with_ddl/down_pd.log /tmp/tidb_cdc_test/kill_owner_with_ddl/cdc.log /tmp/tidb_cdc_test/kill_owner_with_ddl/sync_diff_inspector.log /tmp/tidb_cdc_test/kill_owner_with_ddl/tidb_down.log /tmp/tidb_cdc_test/owner_remove_table_error/tikv2.log /tmp/tidb_cdc_test/owner_remove_table_error/stdout.log /tmp/tidb_cdc_test/owner_remove_table_error/tikv_down/db/000005.log /tmp/tidb_cdc_test/owner_remove_table_error/tikv3.log /tmp/tidb_cdc_test/owner_remove_table_error/tikv_down.log /tmp/tidb_cdc_test/owner_remove_table_error/tikv3/db/000005.log /tmp/tidb_cdc_test/owner_remove_table_error/tidb_other.log /tmp/tidb_cdc_test/owner_remove_table_error/pd1/region-meta/000001.log /tmp/tidb_cdc_test/owner_remove_table_error/pd1/hot-region/000001.log /tmp/tidb_cdc_test/owner_remove_table_error/pd1.log /tmp/tidb_cdc_test/owner_remove_table_error/tikv1.log /tmp/tidb_cdc_test/owner_remove_table_error/cdc_data/tmp/sorter/0002/000002.log /tmp/tidb_cdc_test/owner_remove_table_error/cdc_data/tmp/sorter/0005/000002.log /tmp/tidb_cdc_test/owner_remove_table_error/cdc_data/tmp/sorter/0007/000002.log /tmp/tidb_cdc_test/owner_remove_table_error/cdc_data/tmp/sorter/0001/000002.log /tmp/tidb_cdc_test/owner_remove_table_error/cdc_data/tmp/sorter/0004/000002.log /tmp/tidb_cdc_test/owner_remove_table_error/cdc_data/tmp/sorter/0003/000002.log /tmp/tidb_cdc_test/owner_remove_table_error/cdc_data/tmp/sorter/0000/000002.log /tmp/tidb_cdc_test/owner_remove_table_error/cdc_data/tmp/sorter/0006/000002.log /tmp/tidb_cdc_test/owner_remove_table_error/tikv1/db/000005.log /tmp/tidb_cdc_test/owner_remove_table_error/tikv2/db/000005.log /tmp/tidb_cdc_test/owner_remove_table_error/sync_diff/output/sync_diff.log /tmp/tidb_cdc_test/owner_remove_table_error/tidb.log /tmp/tidb_cdc_test/owner_remove_table_error/tidb-slow.log /tmp/tidb_cdc_test/owner_remove_table_error/down_pd.log /tmp/tidb_cdc_test/owner_remove_table_error/cdc.log /tmp/tidb_cdc_test/owner_remove_table_error/down_pd/region-meta/000001.log /tmp/tidb_cdc_test/owner_remove_table_error/down_pd/hot-region/000001.log /tmp/tidb_cdc_test/owner_remove_table_error/sync_diff_inspector.log /tmp/tidb_cdc_test/owner_remove_table_error/tiflash/db/proxy/db/000005.log /tmp/tidb_cdc_test/owner_remove_table_error/tiflash/log/proxy.log /tmp/tidb_cdc_test/owner_remove_table_error/tiflash/log/server.log /tmp/tidb_cdc_test/owner_remove_table_error/tiflash/log/error.log /tmp/tidb_cdc_test/owner_remove_table_error/tidb_down.log
tar: Removing leading `/' from member names
/tmp/tidb_cdc_test/bank/tikv2.log
/tmp/tidb_cdc_test/default_value/tidb_down.log
/tmp/tidb_cdc_test/default_value/tikv2/db/000005.log
/tmp/tidb_cdc_test/default_value/tidb_other.log
/tmp/tidb_cdc_test/default_value/tidb-slow.log
/tmp/tidb_cdc_test/default_value/down_pd.log
/tmp/tidb_cdc_test/default_value/tikv1/db/000005.log
/tmp/tidb_cdc_test/default_value/down_pd/region-meta/000001.log
/tmp/tidb_cdc_test/default_value/down_pd/hot-region/000001.log
/tmp/tidb_cdc_test/default_value/pd1.log
/tmp/tidb_cdc_test/default_value/tikv_down/db/000005.log
/tmp/tidb_cdc_test/default_value/pd1/region-meta/000001.log
/tmp/tidb_cdc_test/default_value/pd1/hot-region/000001.log
/tmp/tidb_cdc_test/default_value/tidb.log
/tmp/tidb_cdc_test/default_value/tikv3.log
Verifying downstream PD is started...
+ ls -alh log-G10.tar.gz
-rw-r--r-- 1 jenkins jenkins 3.1M May  8 16:06 log-G10.tar.gz
[Pipeline] archiveArtifacts
Archiving artifacts
Starting Upstream TiKV...
TiKV 
Release Version:   8.1.0
Edition:           Community
Git Commit Hash:   56613f7c3e28c02853cc51d15bc1b77f68b58be8
Git Commit Branch: release-8.1
UTC Build Time:    2024-04-30 06:15:40
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Starting Downstream TiKV...
TiKV 
Release Version:   8.1.0
Edition:           Community
Git Commit Hash:   56613f7c3e28c02853cc51d15bc1b77f68b58be8
Git Commit Branch: release-8.1
UTC Build Time:    2024-04-30 06:15:40
Rust Version:      rustc 1.77.0-nightly (89e2160c4 2023-12-27)
Enable Features:   pprof-fp jemalloc mem-profiling portable sse test-engine-kv-rocksdb test-engine-raft-raft-engine trace-async-tasks openssl-vendored
Profile:           dist_release
Recording fingerprints
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // stage
=================>> Running test /home/jenkins/agent/workspace/pingcap/tiflow/release-8.1/pull_cdc_integration_mysql_test/tiflow/tests/integration_tests/kafka_compression/run.sh using Sink-Type: mysql... <<=================
[Wed May  8 16:06:37 CST 2024] <<<<<< run test case kafka_compression success! >>>>>>
[Pipeline] }
/tmp/tidb_cdc_test/bank/stdout.log
/tmp/tidb_cdc_test/bank/tikv3.log
[Pipeline] // container
[Pipeline] }
check diff failed 2-th time, retry later
[Pipeline] // withEnv
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
[Pipeline] }
table ddl_manager.finish_mark not exists for 87-th check, retry later
table consistent_replicate_storage_file.check1 not exists for 16-th check, retry later
[Pipeline] // node
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // stage
/tmp/tidb_cdc_test/bank/tikv_down.log
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G10'
Sending interrupt signal to process
Killing processes
Starting Upstream TiDB...
Release Version: v8.1.0-alpha-107-g238f2e688e
Edition: Community
Git Commit Hash: 238f2e688e9b69c1fa0f167de290828b7d99d172
Git Branch: release-8.1
UTC Build Time: 2024-05-07 09:31:59
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Starting Downstream TiDB...
Release Version: v8.1.0-alpha-107-g238f2e688e
Edition: Community
Git Commit Hash: 238f2e688e9b69c1fa0f167de290828b7d99d172
Git Branch: release-8.1
UTC Build Time: 2024-05-07 09:31:59
GoVersion: go1.21.6
Race Enabled: false
Check Table Before Drop: false
Store: unistore
Verifying Upstream TiDB is started...
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
script returned exit code 143
check diff failed 3-th time, retry later
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
table ddl_manager.finish_mark not exists for 88-th check, retry later
table consistent_replicate_storage_file.check1 not exists for 17-th check, retry later
++ stop_tidb_cluster
{"level":"warn","ts":1715155600.4209445,"caller":"v3@v3.5.12/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0022a88c0/127.0.0.1:2379","attempt":0,"error":"rpc error: code = Unavailable desc = error reading from server: EOF"}
script returned exit code 143
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
{"level":"warn","ts":1715155600.6218724,"caller":"v3@v3.5.12/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0029b1a40/127.0.0.1:2479","attempt":0,"error":"rpc error: code = Unavailable desc = error reading from server: EOF"}
{"level":"warn","ts":1715155600.622762,"caller":"v3@v3.5.12/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0014d3a40/127.0.0.1:2379","attempt":0,"error":"rpc error: code = Unavailable desc = error reading from server: EOF"}
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)

script returned exit code 143
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
\033[0;36m<<< Run all test success >>>\033[0m
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
{"level":"warn","ts":1715155601.1656892,"caller":"v3@v3.5.12/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc002253c00/127.0.0.1:2379","attempt":0,"error":"rpc error: code = Unavailable desc = error reading from server: EOF"}
script returned exit code 143
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
/tmp/tidb_cdc_test/bank/tikv2-2024-05-08T16-03-10.758.log
script returned exit code 143
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
script returned exit code 143
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
kill finished with exit code 0
Sending interrupt signal to process
Killing processes
check diff failed 4-th time, retry later
script returned exit code 143
kill finished with exit code 0
Killing processes
script returned exit code 143
kill finished with exit code 1
Sending interrupt signal to process
Killing processes
{"level":"warn","ts":1715155600.8944821,"caller":"v3@v3.5.12/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc002022a80/127.0.0.1:2379","attempt":0,"error":"rpc error: code = Unavailable desc = error reading from server: EOF"}
script returned exit code 143
/tmp/tidb_cdc_test/bank/tidb_other.log
/tmp/tidb_cdc_test/bank/pd1.log
/tmp/tidb_cdc_test/bank/tikv1.log
kill finished with exit code 0
script returned exit code 143
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (ws/jenkins-pingcap-tiflow-release-8.1-pull_cdc_integration_mysql_test-84/tiflow-cdc already exists)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] }
Cache not saved (inner-step execution failed)
Error when executing failure post condition:
Also:   org.jenkinsci.plugins.workflow.actions.ErrorAction$ErrorId: 7d401e82-3194-4023-8fa9-88b765599d3d
org.jenkinsci.plugins.workflow.steps.FlowInterruptedException
	at org.jenkinsci.plugins.workflow.cps.steps.ParallelStep$ResultHandler$Callback.checkAllDone(ParallelStep.java:151)
	at org.jenkinsci.plugins.workflow.cps.steps.ParallelStep$ResultHandler$Callback.onFailure(ParallelStep.java:138)
	at org.jenkinsci.plugins.workflow.cps.CpsBodyExecution$FailureAdapter.receive(CpsBodyExecution.java:354)
	at com.cloudbees.groovy.cps.impl.ThrowBlock$1.receive(ThrowBlock.java:66)
	at com.cloudbees.groovy.cps.impl.ConstantBlock.eval(ConstantBlock.java:21)
	at com.cloudbees.groovy.cps.Next.step(Next.java:83)
	at com.cloudbees.groovy.cps.Continuable$1.call(Continuable.java:152)
	at com.cloudbees.groovy.cps.Continuable$1.call(Continuable.java:146)
	at org.codehaus.groovy.runtime.GroovyCategorySupport$ThreadCategoryInfo.use(GroovyCategorySupport.java:136)
	at org.codehaus.groovy.runtime.GroovyCategorySupport.use(GroovyCategorySupport.java:275)
	at com.cloudbees.groovy.cps.Continuable.run0(Continuable.java:146)
	at org.jenkinsci.plugins.workflow.cps.SandboxContinuable.access$001(SandboxContinuable.java:18)
	at org.jenkinsci.plugins.workflow.cps.SandboxContinuable.run0(SandboxContinuable.java:51)
	at org.jenkinsci.plugins.workflow.cps.CpsThread.runNextChunk(CpsThread.java:187)
	at org.jenkinsci.plugins.workflow.cps.CpsThreadGroup.run(CpsThreadGroup.java:423)
	at org.jenkinsci.plugins.workflow.cps.CpsThreadGroup$2.call(CpsThreadGroup.java:331)
	at org.jenkinsci.plugins.workflow.cps.CpsThreadGroup$2.call(CpsThreadGroup.java:295)
	at org.jenkinsci.plugins.workflow.cps.CpsVmExecutorService$2.call(CpsVmExecutorService.java:97)
	at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
	at hudson.remoting.SingleLaneExecutorService$1.run(SingleLaneExecutorService.java:139)
	at jenkins.util.ContextResettingExecutorService$1.run(ContextResettingExecutorService.java:28)
	at jenkins.security.ImpersonatingExecutorService$1.run(ImpersonatingExecutorService.java:68)
	at jenkins.util.ErrorLoggingExecutorService.lambda$wrap$0(ErrorLoggingExecutorService.java:51)
	at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
	at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
	at java.base/java.lang.Thread.run(Thread.java:829)

[Pipeline] }
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] // cache
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
script returned exit code 143
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] // dir
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // cache
[Pipeline] }
[Pipeline] // timeout
[Pipeline] // timeout
{"level":"warn","ts":1715155602.970943,"caller":"v3@v3.5.12/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0021b9340/127.0.0.1:2379","attempt":0,"error":"rpc error: code = Unavailable desc = error reading from server: read tcp 127.0.0.1:58478->127.0.0.1:2379: read: connection reset by peer"}
script returned exit code 143
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] // timeout
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // dir
[Pipeline] }
[Pipeline] }
Cache not saved (inner-step execution failed)
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] // cache
[Pipeline] }
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] // container
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // dir
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // withCredentials
[Pipeline] }
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] // node
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G21'
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G00'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G01'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G02'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G04'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G05'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G06'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G07'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G08'
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G09'
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G11'
[Pipeline] // podTemplate
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
Failed in branch Matrix - TEST_GROUP = 'G12'
[Pipeline] // parallel
[Pipeline] }
[Pipeline] // stage
[Pipeline] }
[Pipeline] // timeout
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // container
[Pipeline] }
[Pipeline] // withEnv
[Pipeline] }
[Pipeline] // node
[Pipeline] }
[Pipeline] // podTemplate
[Pipeline] End of Pipeline
ERROR: script returned exit code 1
Finished: FAILURE