You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have a kind of convoluted setup for building boot images for a high performance computing cluster consisting of 4 layers:
a unionfs-fuse over a pristine chroot, a dir with stuff for the image, a dir with stuff with test cases only used during build. This also has a plugin that outputs all accessed files into a log.
the unioned filesystem is then exported via NFS kernel server
kvm with mini initramfs to setup step 4 and pivot_root to unionfs
unionfs over a tmpfs and NFS
The KVM instance boots up and runs a bunch of test cases for all the tools that belong in the boot image and every accessed file is logged by step 1. This gives us a list of files needed in the boot image allowing us to create minimal boot images.
Now the problem is that randomly the test cases get an I/O error. This either causes a Bus Error in an application itself or reading some file fails. This is fatal to ~80% of build attempts at the moment for one specific image and one user as it hits a essential systemd service file. Works fine for another user. Works better when the build server is freshly booted and seem to get slightly worse over time. Something fishy is going on there.
Are there any known random failures with either tmpfs or nfs as branches? Or do you have tips for debugging this without getting a billion lines of strace output?
The text was updated successfully, but these errors were encountered:
...unfortunately, i don't know any known bugs that would be somehow specific to nfs or tmpfs. also, your setup seems too complicated to draw any conclusion or to give better advice than "try to remove some of the layers" (just for testing purposes)... :-(
anyway, if you manage to find the problem and it's really caused by unionfs, i'd love to hear back from you, thanks!
I have a kind of convoluted setup for building boot images for a high performance computing cluster consisting of 4 layers:
The KVM instance boots up and runs a bunch of test cases for all the tools that belong in the boot image and every accessed file is logged by step 1. This gives us a list of files needed in the boot image allowing us to create minimal boot images.
Now the problem is that randomly the test cases get an I/O error. This either causes a Bus Error in an application itself or reading some file fails. This is fatal to ~80% of build attempts at the moment for one specific image and one user as it hits a essential systemd service file. Works fine for another user. Works better when the build server is freshly booted and seem to get slightly worse over time. Something fishy is going on there.
Are there any known random failures with either tmpfs or nfs as branches? Or do you have tips for debugging this without getting a billion lines of strace output?
The text was updated successfully, but these errors were encountered: