Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

A TiKV was restarted before it was stopped for scheduling, and BR did not exit in a timely manner and set the status to failed, resulting in the cluster remaining in the pause schedule state [restore for 5583] #5820

Open
ti-chi-bot opened this issue Oct 25, 2024 · 1 comment
Assignees
Labels
area/ebs-br type/bug Something isn't working

Comments

@ti-chi-bot
Copy link
Member

Bug Report

What version of Kubernetes are you using?

What version of TiDB Operator are you using?

What storage classes exist in the Kubernetes cluster and what are used for PD/TiKV pods?

What's the status of the TiDB cluster pods?

What did you do?

  1. do rolling restart
  2. do volumebackup

What did you expect to see?

  1. volumebackup success
  2. The pause schedule lasts for a short period of time

What did you see instead?

  1. volumebackup is stuck, when exceed Volume Backup Init Job Max Active Seconds, volumebackup is set to failed

img_v3_0295_c15d2c9e-311b-4186-81b8-366bd787a74g

img_v3_0295_7667b5b9-b3a4-4a11-8b60-5cc8996adeeg
img_v3_0294_5b2d712b-d82b-4941-8cdd-1424aa0e540g
image

@ti-chi-bot ti-chi-bot added area/ebs-br type/bug Something isn't working labels Oct 25, 2024
@ti-chi-bot
Copy link
Member Author

This issue is restored for #5583, it originally created at 2024-03-20 03:28:42.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area/ebs-br type/bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants