-
Notifications
You must be signed in to change notification settings - Fork 2.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Migrate CRI-O jobs away from kubernetes_e2e.py
#32567
Comments
/sig node |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
/remove-lifecycle stale |
/triage accepted |
Does this still need help? can i start looking at it? |
@elieser1101 I'd appreciate your eyes on that. 🙏 |
/assign |
@kannon92 @elieser1101 @haircommander
This seems to be a core issue causing jobs to fail. Unfortunately I can't reproduce it in my environment. Here is how I run kubetest2 for splitfs tests:
I suspect that this could be caused by the host/vm kernel and container resource restrictions, but I don't know how to specify upper level instance image (gcr.io/k8s-staging-test-infra/kubekins-e2e:v20241128-8df65c072f-master) and container resources (cpu 4 and memory 6Gi) when running kubetest2 locally. Any ideas how to proceed further? |
do you have access to the nodes you've provisioned @bart0sh ? can I poke around? basically, we want to be able to run |
@haircommander yes, I have access to the nodes, but I can't reproduce the error there :( |
Trigered the job couple of times and seem to be improved(still failed at some point), also takes longer to complete. We could test the imagefs one with the same approach @bart0sh but i guess the right fix includes the selinux cahnge? |
I'm not sure about it. selinux configuration is the same for kubetest2 and old jobs, but only kubetest2 jobs fail. |
decreasing parallelism improved imagefs test. Previously I didn't see any successful job runs. With the change I can see at least one so far. |
I'm still wondering why I can't repro |
Unfortunately using more powerful instance didn't change much for imagefs job. I can still see the same error in the logs. |
@elieser1101 I can see a lot of green kubetest2 jobs in the test grid. Is there anything that prevents replacing kubernetes_e2e.py jobs with them? I did it for splitfs and imagefs jobs as I was involved in fixing them. I can do it for the rest of jobs if needed. |
@bart0sh thank you very much for the splitfs/imagefs that was a great finding What would come next is to validate that the kubetest2 are actually working. Meaning, I noticed that some of the jobs are completing but are skipping all the specs. We would like to ensure we are running the jobs properly before replacing the kubernetes_e2e.py jobs. At the moment im loking at the DRA ones wich were missing some kubetest2 features and this |
The
kubernetes_e2e.py
script is deprecated and we should usekubetest2
instead.All affected tests are listed in https://testgrid.k8s.io/sig-node-cri-o
cc @kubernetes/sig-node-cri-o-test-maintainers
Ref: https://github.com/kubernetes/test-infra/tree/master/scenarios, #20760
The text was updated successfully, but these errors were encountered: