All checks were successful
DCO action / DCO (pull_request) Successful in 38s
Vulncheck / Vulncheck (pull_request) Successful in 55s
Build / Build Components (pull_request) Successful in 1m29s
Pre-commit hooks / Pre-commit (pull_request) Successful in 1m30s
Tests and linters / gopls check (pull_request) Successful in 2m24s
Tests and linters / Run gofumpt (pull_request) Successful in 2m47s
Tests and linters / Tests (pull_request) Successful in 3m1s
Tests and linters / Staticcheck (pull_request) Successful in 3m10s
Tests and linters / Lint (pull_request) Successful in 3m18s
Tests and linters / Tests with -race (pull_request) Successful in 4m0s
There is a race condition when multiple cache operation try to report the cache size metrics simultaneously. Consider the following example: - the initial total size of objects stored in the cache size is 2 - worker X deletes an object and reads the cache size, which is 1 - worker Y deletes an object and reads the cache size, which is 0 - worker Y reports the cache size it learnt, which is 0 - worker X reports the cache size it learnt, which is 1 As a result, the observed cache size is 1 (i. e. one object remains in the cache), which is incorrect because the actual cache size is 0. To fix this, let's report the metrics periodically in the flush loop. Signed-off-by: Aleksey Savchuk <a.savchuk@yadro.com>
20 lines
504 B
Go
20 lines
504 B
Go
package writecache
|
|
|
|
func (c *cache) estimateCacheSize() (uint64, uint64) {
|
|
count, size := c.counter.CountSize()
|
|
c.metrics.SetEstimateSize(size)
|
|
c.metrics.SetActualCounters(count)
|
|
return count, size
|
|
}
|
|
|
|
func (c *cache) hasEnoughSpaceFS() bool {
|
|
return c.hasEnoughSpace(c.maxObjectSize)
|
|
}
|
|
|
|
func (c *cache) hasEnoughSpace(objectSize uint64) bool {
|
|
count, size := c.estimateCacheSize()
|
|
if c.maxCacheCount > 0 && count+1 > c.maxCacheCount {
|
|
return false
|
|
}
|
|
return c.maxCacheSize >= size+objectSize
|
|
}
|