One node stuck in old epoch after shutdown/start all nodes #448
Labels
No labels
P0
P1
P2
P3
badger
frostfs-adm
frostfs-cli
frostfs-ir
frostfs-lens
frostfs-node
good first issue
triage
Infrastructure
blocked
bug
config
discussion
documentation
duplicate
enhancement
go
help wanted
internal
invalid
kludge
observability
perfomance
question
refactoring
wontfix
No project
No assignees
1 participant
Notifications
Due date
No due date set.
Dependencies
No dependencies set.
Reference: TrueCloudLab/frostfs-node#448
Loading…
Reference in a new issue
No description provided.
Delete branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Autotest testsuites.failover.test_failover_full_shutdown.TestFailoverFullShutdown#test_full_storage_shutdown
Expected Behavior
After shutdown/start all nodes should be in the same epoch.
Current Behavior
One node stuck in old epoch after shutdown/start all nodes.
Steps to Reproduce (for bugs)
1.Shutdown all nodes
2.Start all nodes
3.Wait for all nodes will have the following status
4.Created s3 creds:
5.Make preset in k6
6.Start k6 with the following env:
7.k6 finished with the following failed r/w operations:
8.All errors from 10.78.70.138 node:
9.In node log:
This error said that node 138 in old epoch and token has expired.
10.Also node log has the following lines about epoch:
Context
Test objectivity: load should be run before shutdown and after start nodes.
Regression
yes
Version
Your Environment
HW
4 nodes
Datacenter config: 1+1+1+1
Network profile: bad
Bandwidth: 1Gb
Packet loss: 5% с 50% dependency from old loss
Not reproduced on v0.37.0-rc.1-6-g5acc13fa
Suggest to close.