TiCDC 新架构v9.0.0 panic且无法启动

【TiDB 使用环境】测试
【TiDB 版本】
TiDB v8.5.1 单机集群, TiCDC升级到 v9.0.0

【操作系统】
rocky9.5
128G MEM, 16CPU, SSD

【部署方式】
单机集群

【问题复现路径】做过哪些操作出现的问题
配置完成TiCDC,TiDB 8.5.1 到mysql5.7.40的数据同步。
小批量数据同步测试正常。
大批量数据加载同步过程中报错。:
# tiup bench tpcc -H192.168.169.40 -P4000 -D tpcc -Uroot -ptidb --warehouses 40 --parts 4 prepare

【遇到的问题:问题现象及影响】
发现TiCDC down了,随后刷新集群状态TiCDC 在Up和Down间切换,数次后彻底为down状态。
重启TiDB 集群,显示TiCDC启动成功,但看状态还是为done.

[root@tidb40 log]# tiup cluster start tidb-test
Starting cluster tidb-test...
+ [ Serial ] - SSHKeySet: privateKey=/root/.tiup/storage/cluster/clusters/tidb-test/ssh/id_rsa, publicKey=/root/.tiup/storage/cluster/clusters/tidb-test/ssh/id_rsa.pub
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [ Serial ] - StartCluster
Starting component pd
        Starting instance 192.168.169.40:2379
        Start instance 192.168.169.40:2379 success
Starting component tikv
        Starting instance 192.168.169.40:20162
        Starting instance 192.168.169.40:20160
        Starting instance 192.168.169.40:20161
        Start instance 192.168.169.40:20162 success
        Start instance 192.168.169.40:20160 success
        Start instance 192.168.169.40:20161 success
Starting component tidb
        Starting instance 192.168.169.40:4000
        Start instance 192.168.169.40:4000 success
Starting component tiflash
        Starting instance 192.168.169.40:9000
        Start instance 192.168.169.40:9000 success
Starting component cdc
        Starting instance 192.168.169.40:8302
        Starting instance 192.168.169.40:8300
        Starting instance 192.168.169.40:8301
        Start instance 192.168.169.40:8302 success
        Start instance 192.168.169.40:8300 success
        Start instance 192.168.169.40:8301 success
Starting component prometheus
        Starting instance 192.168.169.40:9090
        Start instance 192.168.169.40:9090 success
Starting component grafana
        Starting instance 192.168.169.40:3000
        Start instance 192.168.169.40:3000 success
Starting component node_exporter
        Starting instance 192.168.169.40
        Start 192.168.169.40 success
Starting component blackbox_exporter
        Starting instance 192.168.169.40
        Start 192.168.169.40 success
+ [ Serial ] - UpdateTopology: cluster=tidb-test
Started cluster `tidb-test` successfully

【复制黏贴 ERROR 报错的日志】
CDC 的错误日志显示panic

[root@tidb40 log]# pwd
/tidb-deploy/cdc-8302/log
[root@tidb40 log]# more cdc_stderr.log
panic: should not reach here

goroutine 932 [running]:
go.uber.org/zap/zapcore.CheckWriteAction.OnWrite(0x1?, 0x1?, {0x0?, 0x0?, 0xc00b0ceac0?})
        go.uber.org/zap@v1.27.0/zapcore/entry.go:196 +0x54
go.uber.org/zap/zapcore.(*CheckedEntry).Write(0xc0144420d0, {0xc0058f5340, 0x1, 0x1})
        go.uber.org/zap@v1.27.0/zapcore/entry.go:262 +0x24e
go.uber.org/zap.(*Logger).Panic(0x0?, {0x58e273d?, 0xc0058f5340?}, {0xc0058f5340, 0x1, 0x1})
        go.uber.org/zap@v1.27.0/logger.go:285 +0x51
github.com/pingcap/log.Panic({0x58e273d?, 0x5726f40?}, {0xc0058f5340?, 0x407838?, 0x55b11b?})
        github.com/pingcap/log@v1.1.1-0.20241212030209-7e3ff8601a2a/global.go:54 +0x85
github.com/pingcap/ticdc/logservice/schemastore.extractTableInfoFuncForDropTable(0x4ea1fc0?, 0xc003499aa0?)
        github.com/pingcap/ticdc/logservice/schemastore/persist_storage_ddl_handlers.go:1352 +0xc9
github.com/pingcap/ticdc/logservice/schemastore.(*versionedTableInfoStore).doApplyDDL(0xc008de6960, 0xc00dd8e380)
        github.com/pingcap/ticdc/logservice/schemastore/multi_version.go:205 +0x5ad
github.com/pingcap/ticdc/logservice/schemastore.(*versionedTableInfoStore).applyDDL(0xc008de6960, 0xc003b7e360?)
        github.com/pingcap/ticdc/logservice/schemastore/multi_version.go:187 +0x1b6
github.com/pingcap/ticdc/logservice/schemastore.(*persistentStorage).handleDDLJob.func1({0xc005922640, 0x4, 0xc00557b818?})
        github.com/pingcap/ticdc/logservice/schemastore/persist_storage.go:704 +0x19e
github.com/pingcap/ticdc/logservice/schemastore.iterateEventTablesForSingleTableDDL(0xc00dd8e380, 0xc00b0ceb60)
        github.com/pingcap/ticdc/logservice/schemastore/persist_storage_ddl_handlers.go:1208 +0x16e
github.com/pingcap/ticdc/logservice/schemastore.(*persistentStorage).handleDDLJob(0xc0045022c0, 0xc00019a4e0)
        github.com/pingcap/ticdc/logservice/schemastore/persist_storage.go:694 +0x3dd
github.com/pingcap/ticdc/logservice/schemastore.(*schemaStore).updateResolvedTsPeriodically.func1()
        github.com/pingcap/ticdc/logservice/schemastore/schema_store.go:210 +0x133b
github.com/pingcap/ticdc/logservice/schemastore.(*schemaStore).updateResolvedTsPeriodically(0xc00413c2d0, {0x63cc6a0, 0xc0036ea230})
        github.com/pingcap/ticdc/logservice/schemastore/schema_store.go:231 +0xe8
github.com/pingcap/ticdc/logservice/schemastore.(*schemaStore).Run.func2()
        github.com/pingcap/ticdc/logservice/schemastore/schema_store.go:140 +0x1f
golang.org/x/sync/errgroup.(*Group).Go.func1()
        golang.org/x/sync@v0.10.0/errgroup/errgroup.go:78 +0x50
created by golang.org/x/sync/errgroup.(*Group).Go in goroutine 1035
        golang.org/x/sync@v0.10.0/errgroup/errgroup.go:75 +0x96
panic: should not reach here

3个节点的TiCDC 日志和报错日志如下:
8302cdc_stderr.zip (5.7 KB)
8302cdc.zip (739.8 KB)
8301cdc.zip (923.3 KB)
8300cdc_stderr.zip (6.6 KB)
8300cdc.zip (839.9 KB)
8301cdc_stderr.zip (7.0 KB)

【其他附件:截图/日志/监控】






已经排查到原因,会尽快修复,修复完会在这里同步。

2 个赞

问题已修复,有空可以用最新的 nightly 试一下

怎么操作?
看原地址:https://tiup-mirrors.pingcap.com/cdc-v9.0.0-alpha-nightly-linux-amd64.tar.gz 文件没变。


[root@tidb40 soft]# md5sum cdc-v9.0.0-alpha-nightly-linux-amd64*
0b1d04c2bef78383dfc342c11f2ed34e  cdc-v9.0.0-alpha-nightly-linux-amd64_new.tar.gz
0b1d04c2bef78383dfc342c11f2ed34e  cdc-v9.0.0-alpha-nightly-linux-amd64.tar.gz

1 个赞

https://tiup-mirrors.pingcap.com/cdc-v9.0.0-beta.1.pre-nightly-linux-amd64.tar.gz 用这个地址下载最新的 nightly 版本,之前文档里的链接会尽快更新。

刚在原单机集群环境升级cdc失败。

怀疑:新cdc需要在8.5.1环境上升级,不能在原9.0.0环境上升级。

升级前集群状态正常;

cdc从9.0.0原版本升级新版本失败;


[root@tidb40 soft]# tiup cluster patch tidb-test cdc-v9.0.0-beta.1.pre-nightly-linux-arm64.tar.gz -R cdc --overwrite
Will patch the cluster tidb-test with package path is cdc-v9.0.0-beta.1.pre-nightly-linux-arm64.tar.gz, nodes: , roles: cdc.
Do you want to continue? [y/N]:(default=N) y
+ [ Serial ] - SSHKeySet: privateKey=/root/.tiup/storage/cluster/clusters/tidb-test/ssh/id_rsa, publicKey=/root/.tiup/storage/cluster/clusters/tidb-test/ssh/id_rsa.pub
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [ Serial ] - BackupComponent: component=cdc, currentVersion=v8.5.1, remote=192.168.169.40:/tidb-deploy/cdc-8302
+ [ Serial ] - BackupComponent: component=cdc, currentVersion=v8.5.1, remote=192.168.169.40:/tidb-deploy/cdc-8300
+ [ Serial ] - BackupComponent: component=cdc, currentVersion=v8.5.1, remote=192.168.169.40:/tidb-deploy/cdc-8301
+ [ Serial ] - InstallPackage: srcPath=cdc-v9.0.0-beta.1.pre-nightly-linux-arm64.tar.gz, remote=192.168.169.40:/tidb-deploy/cdc-8302
+ [ Serial ] - InstallPackage: srcPath=cdc-v9.0.0-beta.1.pre-nightly-linux-arm64.tar.gz, remote=192.168.169.40:/tidb-deploy/cdc-8300
+ [ Serial ] - InstallPackage: srcPath=cdc-v9.0.0-beta.1.pre-nightly-linux-arm64.tar.gz, remote=192.168.169.40:/tidb-deploy/cdc-8301
+ [ Serial ] - UpgradeCluster
Upgrading component cdc
         Start drain the capture, address: 192.168.169.40:8300, captureID: 20db13c6-5279-404f-a1f2-c89264786b95
        Restarting instance 192.168.169.40:8300

Error: failed to restart: 192.168.169.40 cdc-8300.service, please check the instance's log(/tidb-deploy/cdc-8300/log) for more detail.: timed out waiting for port 8300 to be started after 2m0s

Verbose debug logs has been written to /root/.tiup/logs/tiup-cluster-debug-2025-03-18-18-40-09.log.

cdc-8300 节点的cdc_stderr.log

/tidb-deploy/cdc-8300/scripts/run_cdc.sh: line 8: /tidb-deploy/cdc-8300/bin/cdc: cannot execute binary file: Exec format error
/tidb-deploy/cdc-8300/scripts/run_cdc.sh: line 8: /tidb-deploy/cdc-8300/bin/cdc: Success

cdc文件不可执行

[root@tidb40 bin]# ls -al
total 239340
drwxr-xr-x 2 tidb tidb        17 Mar 18 18:38 .
drwxr-xr-x 7 tidb tidb        77 Mar 17 14:00 ..
-rwxr-xr-x 1 tidb tidb 245080739 Mar 18 14:05 cdc
[root@tidb40 bin]# ./cdc -V
-bash: ./cdc: cannot execute binary file: Exec format error

重启tidb集群,cdc节点重启失败,提示超时。

抱歉,上面我的链接贴错了,发了 arm 架构的二进制链接,现在更正了一下,有空麻烦再试试

升级后,无法启动CDC,日志提示"panic: get table info from schemaStore failed"

  1. 升级成功
[root@tidb40 soft]# tiup cluster patch tidb-test cdc-v9.0.0-beta.1.pre-nightly-linux-amd64.tar.gz -R cdc --overwrite
Will patch the cluster tidb-test with package path is cdc-v9.0.0-beta.1.pre-nightly-linux-amd64.tar.gz, nodes: , roles: cdc.
Do you want to continue? [y/N]:(default=N) y
+ [ Serial ] - SSHKeySet: privateKey=/root/.tiup/storage/cluster/clusters/tidb-test/ssh/id_rsa, publicKey=/root/.tiup/storage/cluster/c                                      lusters/tidb-test/ssh/id_rsa.pub
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [ Serial ] - BackupComponent: component=cdc, currentVersion=v8.5.1, remote=192.168.169.40:/tidb-deploy/cdc-8302
+ [ Serial ] - BackupComponent: component=cdc, currentVersion=v8.5.1, remote=192.168.169.40:/tidb-deploy/cdc-8301
+ [ Serial ] - BackupComponent: component=cdc, currentVersion=v8.5.1, remote=192.168.169.40:/tidb-deploy/cdc-8300
+ [ Serial ] - InstallPackage: srcPath=cdc-v9.0.0-beta.1.pre-nightly-linux-amd64.tar.gz, remote=192.168.169.40:/tidb-deploy/cdc-8300
+ [ Serial ] - InstallPackage: srcPath=cdc-v9.0.0-beta.1.pre-nightly-linux-amd64.tar.gz, remote=192.168.169.40:/tidb-deploy/cdc-8301
+ [ Serial ] - InstallPackage: srcPath=cdc-v9.0.0-beta.1.pre-nightly-linux-amd64.tar.gz, remote=192.168.169.40:/tidb-deploy/cdc-8302
+ [ Serial ] - UpgradeCluster
Upgrading component cdc
        Restarting instance 192.168.169.40:8300
        Restart instance 192.168.169.40:8300 success
         Start drain the capture, address: 192.168.169.40:8301, captureID: a7ce2142-79ce-4a03-9aae-487c646f32a9
        Restarting instance 192.168.169.40:8301
        Restart instance 192.168.169.40:8301 success
         Start drain the capture, address: 192.168.169.40:8302, captureID: e07b2d8c-d725-4992-85a7-0b13009c1ad5
        Restarting instance 192.168.169.40:8302
        Restart instance 192.168.169.40:8302 success
Stopping component node_exporter
        Stopping instance 192.168.169.40
        Stop 192.168.169.40 success
Stopping component blackbox_exporter
        Stopping instance 192.168.169.40
        Stop 192.168.169.40 success
Starting component node_exporter
        Starting instance 192.168.169.40
        Start 192.168.169.40 success
Starting component blackbox_exporter
        Starting instance 192.168.169.40
        Start 192.168.169.40 success

2.服务处于关闭状态:

3.重启集群,但CDC依然down;


[root@tidb40 soft]# tiup cluster restart tidb-test
Will restart the cluster tidb-test with nodes: all roles: all.
Cluster will be unavailable
Do you want to continue? [y/N]:(default=N) y
+ [ Serial ] - SSHKeySet: privateKey=/root/.tiup/storage/cluster/clusters/tidb-test/ssh/id_rsa, publicKey=/root/.tiup/storage/cluster/clusters/tidb-test/ssh/id_rsa.pub
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [Parallel] - UserSSH: user=tidb, host=192.168.169.40
+ [ Serial ] - RestartCluster
Stopping component grafana
        Stopping instance 192.168.169.40
        Stop grafana 192.168.169.40:3000 success
Stopping component prometheus
        Stopping instance 192.168.169.40
        Stop prometheus 192.168.169.40:9090 success
Stopping component cdc
        Stopping instance 192.168.169.40
        Stop cdc 192.168.169.40:8300 success
        Stopping instance 192.168.169.40
        Stop cdc 192.168.169.40:8301 success
        Stopping instance 192.168.169.40
        Stop cdc 192.168.169.40:8302 success
Stopping component tiflash
        Stopping instance 192.168.169.40
        Stop tiflash 192.168.169.40:9000 success
Stopping component tidb
        Stopping instance 192.168.169.40
        Stop tidb 192.168.169.40:4000 success
Stopping component tikv
        Stopping instance 192.168.169.40
        Stopping instance 192.168.169.40
        Stopping instance 192.168.169.40
        Stop tikv 192.168.169.40:20160 success
        Stop tikv 192.168.169.40:20161 success
        Stop tikv 192.168.169.40:20162 success
Stopping component pd
        Stopping instance 192.168.169.40
        Stop pd 192.168.169.40:2379 success
Stopping component node_exporter
        Stopping instance 192.168.169.40
        Stop 192.168.169.40 success
Stopping component blackbox_exporter
        Stopping instance 192.168.169.40
        Stop 192.168.169.40 success
Starting component pd
        Starting instance 192.168.169.40:2379
        Start instance 192.168.169.40:2379 success
Starting component tikv
        Starting instance 192.168.169.40:20162
        Starting instance 192.168.169.40:20160
        Starting instance 192.168.169.40:20161
        Start instance 192.168.169.40:20161 success
        Start instance 192.168.169.40:20160 success
        Start instance 192.168.169.40:20162 success
Starting component tidb
        Starting instance 192.168.169.40:4000
        Start instance 192.168.169.40:4000 success
Starting component tiflash
        Starting instance 192.168.169.40:9000
        Start instance 192.168.169.40:9000 success
Starting component cdc
        Starting instance 192.168.169.40:8300
        Starting instance 192.168.169.40:8301
        Starting instance 192.168.169.40:8302
        Start instance 192.168.169.40:8301 success
        Start instance 192.168.169.40:8300 success
        Start instance 192.168.169.40:8302 success
Starting component prometheus
        Starting instance 192.168.169.40:9090
        Start instance 192.168.169.40:9090 success
Starting component grafana
        Starting instance 192.168.169.40:3000
        Start instance 192.168.169.40:3000 success
Starting component node_exporter
        Starting instance 192.168.169.40
        Start 192.168.169.40 success
Starting component blackbox_exporter
        Starting instance 192.168.169.40
        Start 192.168.169.40 success
Restarted cluster `tidb-test` successfully
[root@tidb40 soft]# tiup cluster display tidb-test
Cluster type:       tidb
Cluster name:       tidb-test
Cluster version:    v8.5.1
Deploy user:        tidb
SSH type:           builtin
Dashboard URL:      http://192.168.169.40:2379/dashboard
Grafana URL:        http://192.168.169.40:3000
ID                    Role           Host            Ports                            OS/Arch       Status   Data Dir                    Deploy Dir
--                    ----           ----            -----                            -------       ------   --------                    ----------
192.168.169.40:8300   cdc (patched)  192.168.169.40  8300                             linux/x86_64  Down     /tidb-data/cdc-8300         /tidb-deploy/cdc-8300
192.168.169.40:8301   cdc (patched)  192.168.169.40  8301                             linux/x86_64  Down     /tidb-data/cdc-8301         /tidb-deploy/cdc-8301
192.168.169.40:8302   cdc (patched)  192.168.169.40  8302                             linux/x86_64  Down     /tidb-data/cdc-8302         /tidb-deploy/cdc-8302
192.168.169.40:3000   grafana        192.168.169.40  3000                             linux/x86_64  Up       -                           /tidb-deploy/grafana-3000
192.168.169.40:2379   pd             192.168.169.40  2379/2380                        linux/x86_64  Up|L|UI  /tidb-data/pd-2379          /tidb-deploy/pd-2379
192.168.169.40:9090   prometheus     192.168.169.40  9090/12020                       linux/x86_64  Up       /tidb-data/prometheus-9090  /tidb-deploy/prometheus-9090
192.168.169.40:4000   tidb           192.168.169.40  4000/10080                       linux/x86_64  Up       -                           /tidb-deploy/tidb-4000
192.168.169.40:9000   tiflash        192.168.169.40  9000/3930/20170/20292/8234/8123  linux/x86_64  Up       /tidb-data/tiflash-9000     /tidb-deploy/tiflash-9000
192.168.169.40:20160  tikv           192.168.169.40  20160/20180                      linux/x86_64  Up       /tidb-data/tikv-20160       /tidb-deploy/tikv-20160
192.168.169.40:20161  tikv           192.168.169.40  20161/20181                      linux/x86_64  Up       /tidb-data/tikv-20161       /tidb-deploy/tikv-20161
192.168.169.40:20162  tikv           192.168.169.40  20162/20182                      linux/x86_64  Up       /tidb-data/tikv-20162       /tidb-deploy/tikv-20162
Total nodes: 11
[root@tidb40 soft]#

查看cdc 版本;


[root@tidb40 bin]# cdc version
Release Version: v9.0.0-beta.1.pre-166-g6667aec
Git Commit Hash: 6667aec2dfb5cf2018df3c665b4440aad44779de
Git Branch: HEAD
UTC Build Time: 2025-03-18 06:03:25
Go Version: go1.23.7
Failpoint Build: false

cdc_stderr.log日志 提示如下类似错误:


goroutine 1022 [running]:
go.uber.org/zap/zapcore.CheckWriteAction.OnWrite(0x3?, 0x3?, {0x0?, 0x0?, 0xc003402960?})
        go.uber.org/zap@v1.27.0/zapcore/entry.go:196 +0x54
go.uber.org/zap/zapcore.(*CheckedEntry).Write(0xc003522f70, {0xc003539800, 0x3, 0x3})
        go.uber.org/zap@v1.27.0/zapcore/entry.go:262 +0x24e
go.uber.org/zap.(*Logger).Panic(0x0?, {0x5978c7a?, 0xc003539800?}, {0xc003539800, 0x3, 0x3})
        go.uber.org/zap@v1.27.0/logger.go:285 +0x51
github.com/pingcap/log.Panic({0x5978c7a?, 0xaa?}, {0xc003539800?, 0xc003271780?, 0x65667c77dbc0008?})
        github.com/pingcap/log@v1.1.1-0.20241212030209-7e3ff8601a2a/global.go:54 +0x85
github.com/pingcap/ticdc/pkg/eventservice.(*eventBroker).addDispatcher(0xc0016e3a40, {0x641b0c0, 0xc00351c8e8})
        github.com/pingcap/ticdc/pkg/eventservice/event_broker.go:868 +0x11f9
github.com/pingcap/ticdc/pkg/eventservice.(*eventService).registerDispatcher(0xc0041b03c0, {0x63e0320, 0xc0041b05a0}, {0x641b0c0, 0xc00351c8e8})
        github.com/pingcap/ticdc/pkg/eventservice/event_service.go:145 +0xaf
github.com/pingcap/ticdc/pkg/eventservice.(*eventService).Run(0xc0041b03c0, {0x63e0320, 0xc0041b05a0})
        github.com/pingcap/ticdc/pkg/eventservice/event_service.go:97 +0x191
github.com/pingcap/ticdc/server.(*server).Run.(*server).Run.func2.func3()
        github.com/pingcap/ticdc/server/server.go:237 +0x2fa
golang.org/x/sync/errgroup.(*Group).Go.func1()
        golang.org/x/sync@v0.10.0/errgroup/errgroup.go:78 +0x50
created by golang.org/x/sync/errgroup.(*Group).Go in goroutine 1
        golang.org/x/sync@v0.10.0/errgroup/errgroup.go:75 +0x96
panic: get table info from schemaStore failed

goroutine 1000 [running]:
go.uber.org/zap/zapcore.CheckWriteAction.OnWrite(0x3?, 0x3?, {0x0?, 0x0?, 0xc00374ede0?})
        go.uber.org/zap@v1.27.0/zapcore/entry.go:196 +0x54
go.uber.org/zap/zapcore.(*CheckedEntry).Write(0xc00348c340, {0xc0052ed980, 0x3, 0x3})
        go.uber.org/zap@v1.27.0/zapcore/entry.go:262 +0x24e
go.uber.org/zap.(*Logger).Panic(0x0?, {0x5978c7a?, 0x0?}, {0xc0052ed980, 0x3, 0x3})
        go.uber.org/zap@v1.27.0/logger.go:285 +0x51
github.com/pingcap/log.Panic({0x5978c7a?, 0xa4?}, {0xc0052ed980?, 0xc004ee7c00?, 0x65667c77dbc0008?})
        github.com/pingcap/log@v1.1.1-0.20241212030209-7e3ff8601a2a/global.go:54 +0x85
github.com/pingcap/ticdc/pkg/eventservice.(*eventBroker).addDispatcher(0xc00158cf00, {0x641b0c0, 0xc004045860})
        github.com/pingcap/ticdc/pkg/eventservice/event_broker.go:868 +0x11f9
github.com/pingcap/ticdc/pkg/eventservice.(*eventService).registerDispatcher(0xc004304460, {0x63e0320, 0xc004304640}, {0x641b0c0, 0xc004045860})
        github.com/pingcap/ticdc/pkg/eventservice/event_service.go:145 +0xaf
github.com/pingcap/ticdc/pkg/eventservice.(*eventService).Run(0xc004304460, {0x63e0320, 0xc004304640})
        github.com/pingcap/ticdc/pkg/eventservice/event_service.go:97 +0x191
github.com/pingcap/ticdc/server.(*server).Run.(*server).Run.func2.func3()
        github.com/pingcap/ticdc/server/server.go:237 +0x2fa
golang.org/x/sync/errgroup.(*Group).Go.func1()
        golang.org/x/sync@v0.10.0/errgroup/errgroup.go:78 +0x50
created by golang.org/x/sync/errgroup.(*Group).Go in goroutine 1
        golang.org/x/sync@v0.10.0/errgroup/errgroup.go:75 +0x96
panic: get table info from schemaStore failed

@lidezhu-PingCAP 德竹老师有时间再看看呐。

麻烦把完整的 cdc 日志上传一下

3个CDC 节点日志如下,供分析。
cdc8300_Log.tar.gz (20.0 MB)
cdc8301_Log.tar.gz (24.7 MB)
cdc8302_Log.tar.gz (37.3 MB)

如官方环境测试无问题,可忽略这种情况,即从cdc-v9.0.0-alpha-nightly-linux-amd64.tar.gz 升级到cdc-v9.0.0-beta.1.pre-nightly-linux-amd64.tar.gz,也许是个小概率特殊情况。

收到,我们先分析一下

问题已经定位,会尽快修复,目前的版本重新用当前时间创建一下 changefeed 应该可以绕过

正好有这么一个特殊环境,为解决bug做一点点贡献……
不急,这会不需要用这功能。

CDC 都启动不起来,咋创建?

好的,不着急的话就等一下修复吧

问题已修复,有空可以用最新的 Nightly 再试一下