Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Test failure] Active Defrag on large item: cluster #1454

Open
madolson opened this issue Dec 18, 2024 · 4 comments · May be fixed by #1492
Open

[Test failure] Active Defrag on large item: cluster #1454

madolson opened this issue Dec 18, 2024 · 4 comments · May be fixed by #1492
Assignees
Labels
test-failure An issue indicating a test failure

Comments

@madolson
Copy link
Member

There seems to be one active defrag issue left, where the defragmentation is around 1.1 - 1.2 when it is supposed to be less than 1.1. I am at a bit of a loss for this one, I wasn't able to figure out any local reproduction.

Some example:

  1. https://github.com/valkey-io/valkey/actions/runs/12363680454/job/34505495158#step:8:6531
  2. https://github.com/valkey-io/valkey/actions/runs/12363680454/job/34505490704#step:8:6457
  3. https://github.com/valkey-io/valkey/actions/runs/12363680454/job/34505492618#step:8:6515

@zvi-code @JimB123 Any thoughts?

@madolson madolson added the test-failure An issue indicating a test failure label Dec 18, 2024
@JimB123
Copy link
Contributor

JimB123 commented Dec 18, 2024

In some of these tests, memory usage is low. We set minimum frag to 1.1, but when the memory is small, I think that we can fail to achieve this. The tests don't all have minimum frag levels which would ensure that 1.1 is meaningful. I saw locally a case where defrag just kept spinning and couldn't actually achieve 1.1. Another possibility is that it IS achieving 1.1, and defrag shuts off. But then a little new fragmentation is created during the INFO command which just marginally increases it again. Need to make sure that the success level is a little above the lower frag level to account for this.

I looked earlier this week to see if there were any open test issues, and didn't see any. Looks like we still have a few flaky ones.

I can dive deeper into any flaky ones after the holiday. If this is critical to eliminate the flaky tests now, I'd recommend just adjusting the success criteria a bit until we can do a deeper analysis.

@JimB123
Copy link
Contributor

JimB123 commented Dec 18, 2024

Also, the one for "edge case" is a little weird. I have an open question to @zvi-code to get a second opinion. I'm not sure that the test is meaningful and maybe it should just be eliminated.

@madolson
Copy link
Member Author

I can dive deeper into any flaky ones after the holiday. If this is critical to eliminate the flaky tests now, I'd recommend just adjusting the success criteria a bit until we can do a deeper analysis.

I wouldn't say it's critical to eliminate, I would just like the project to get to a healthier state after the holidays.

@madolson
Copy link
Member Author

I might have a fix with #1492

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
test-failure An issue indicating a test failure
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants