-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
infrequent docker exec .... signal: broken pipe
during kubeadm init | kubeadm join
#949
Comments
per the stack trace, this is definitely failing in currently I've exclusively found these with however in this example: https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-kind-conformance-ipv6/1182717569350504448 you can also see everything else I can find involves |
another example kubernetes/kubernetes#83956 (comment) |
our docker is fairly old, kubernetes/test-infra#14784 |
Still trying to get us on newer docker. Couple of pending test-infra PRs. |
not much progress today -- infra wg this morning and then mitigating, debugging, and dealing with kubernetes/test-infra#14812 now have kubernetes/test-infra#14820 to see about using newer docker in slightly streamlined kind CI specific image experimentally. |
here's an example with
|
have kind experimentally on docker 19.03.X on debian buster. will follow up in the morning. |
docker exec .... signal: broken pipe
during kubeadm initdocker exec .... signal: broken pipe
during kubeadm init | kubeadm join
~everything in kubernetes CI should be on 19.03.X now, we'll have to wait to see if we continue to get these |
still an issue with new docker, https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-kind-conformance-parallel-ipv6/1185195854433095681 here's one with join. |
did test-infra move to 19.03 already? |
I moved the test-infra images, yes. |
ok, thanks for letting me know. but i guess we should add it as "verified" soon. |
@neolit123 we're not running kubeadm / docker against that. the dind image is on 19.03 (kubekins-e2e, and kind's KRTE), but the hosts are on whatever the hosts are on, and the kind nodes are on whatever the kind nodes run |
too many layers :-) |
indeed. :) just wondering when... kinder doesn't have kind node images with docker 19.03 yet, maybe that's the switching point for |
@neolit123 I wouldn't read much into us using 19.03 to host the nodes, someday it might be podman or ignite, it won't reflect much on qualifying with kubeadm. Mostly a shot in the dark regarding the stability issues. I think there's a small chance the root cause of #971 is related here, the go program would get a broken pipe signal if the internal pipe is closed after the internal |
We haven't had one since the patch for #971 went in. However I want to wait a bit longer before calling this fixed. |
we are however seeing this with
hopefully unrelated, need to investigate. EDIT: traced the code, we prefer returning the error from the process, versus from the reader, so it's likely we're seeing this because the reader errored, which would not surprise me for the current untar routine... filed #992 to debug |
we've not had any of these creation failures since #971, granted it has not been an extremely large amount of time. tentatively closing, but still monitoring. will file a new issue for untar issues they don't appear to be related. |
still haven't identified another one since that fix. |
still no signs of this, I think we're in the clear on this one. |
so far I've not once seen this locally, but we have this occasionally in CI, maybe a few times a day with many, many runs.
/assign
/lifecycle active
Examples
https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/83914/pull-kubernetes-e2e-kind/1183813161560576002
#928 (comment)
kubernetes/kubernetes#83903
The text was updated successfully, but these errors were encountered: