tispark报cop task错误

tidb: 5.2.3 ,tispark 2.4.3

spark-sql执行一个count报错如下:
spark-sql>
> select count(*) from ti_o_credit_stop_open_wp_4;
22/12/28 10:23:57 WARN RegionStoreClient: Other error occurred, message: Coprocessor task terminated due to exceeding the deadline
22/12/28 10:23:57 ERROR DAGIterator: Process region tasks failed, remain 0 tasks not executed due to
com.pingcap.tikv.exception.GrpcException: Coprocessor task terminated due to exceeding the deadline
at com.pingcap.tikv.region.RegionStoreClient.handleCopResponse(RegionStoreClient.java:717)
at com.pingcap.tikv.region.RegionStoreClient.coprocess(RegionStoreClient.java:664)
at com.pingcap.tikv.operation.iterator.DAGIterator.process(DAGIterator.java:219)
at com.pingcap.tikv.operation.iterator.DAGIterator.lambda$submitTasks$1(DAGIterator.java:90)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
报错后没影响结果输出。集群配置如下
tispark_masters:

  • host: xxx.146
    ssh_port: 22
    port: 7077
    web_port: 8080
    deploy_dir: /data/tispark/master
    spark_config:
    spark.driver.memory: 8g
    spark_env:
    SPARK_EXECUTOR_CORES: “12”
    SPARK_EXECUTOR_MEMORY: 12g
    SPARK_WORKER_CORES: “12”
    SPARK_WORKER_MEMORY: 12g
    arch: arm64
    os: linux
    tispark_workers:
  • host: xx.146
    ssh_port: 22
    port: 7078
    web_port: 8081
    deploy_dir: /data/tispark/worker/deploy
    arch: arm64
    os: linux
  • host: xxx.151
    ssh_port: 22
    port: 7078
    web_port: 8081
    deploy_dir: /data/tispark/worker/deploy
    arch: arm64
    os: linux
  • host: xx.152
    ssh_port: 22
    port: 7078
    web_port: 8081
    deploy_dir: /data/tispark/worker/deploy
    arch: arm64
    os: linux

可以考虑配置下 超时时间。 spark.tispark.grpc.timeout_in_sec

master/worker的spark-defaults.conf都添加了 ,调整到了300秒,还是报错

这个参数好使,前面没生效是因为修改参数后用的tiup reload方式重启的导致新添加的参数被删除了。用restart方式可以

此话题已在最后回复的 60 天后被自动关闭。不再允许新回复。