-
Notifications
You must be signed in to change notification settings - Fork 10
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ownership error with squashed outputs #142
Comments
@fcomitani — Thank you for the issue submission. A few questions if you could elucidate:
I'm glad you were able to find a workaround in the interim! @rcurrie — Do you know if this is a common setup for the workflow for our Treehouse collaborators? |
The Treehouse collaborators are all running using docker to ensure concordance as the output is added to the public compendium. |
@rcurrie — Thanks for the info. To get around Docker permission issues are there any groups trying to run Docker in a VM like Federico or has that not come up as an issue? |
@jvivian Doh! My bad, didn't connect the dots here. I'm 99% sure BC and Nationwide both run Docker on the host OS vs. in a VM. I just unpacked a Nationwide tar and it seems to be find (files appear as created/owned by me) |
@jvivian thanks for getting back to me.
Thanks again. |
Hello,
I just wanted to notify an issue that arises when dealing with a change of ownership in the pipeline output files.
I am running the pipeline as root on a dedicated VM within an HPC cluster for which all output is squashed to a specific user. This workaround was set up to avoid security issues with docker on the HPC cluster.
The problem is due to tar, which requires to explicitly acknowledge the change in ownership.
I managed to solve the problem first by adding
--no-same-owner
to every tar call in tools/aligners.py, preprocessing.py and quantifiers.py as in the following example:subprocess.check_call(['tar', '-xvf', os.path.join(job.tempDir, 'starIndex.tar.gz'), '-C', job.tempDir, '--no-same-owner'])
To temporarily bypass the problem when creating the output tarball files, however, I had to explicitly add the username information to the code of utils/files.py.
Line 20
f_out.add(file_path, arcname=arcname)
is nowI'm sure there are more elegant solutions to this, but I wanted to let you know in case anybody else tries to run the pipeline on a VM.
Regards,
Federico
The text was updated successfully, but these errors were encountered: