-
Notifications
You must be signed in to change notification settings - Fork 38
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
rhel 8.10 - compute-redhat.yml breaks image build #419
Comments
A dump of found 'issues'
Adding a print statement to python like so
|
"osimage pack for compute already queued" normally means that another packing for that image was already in progress. It prevents it from being packed twice at the same time. However if changes were made while the other packing was already in progress, things will go wrong. Was there only one packing active at that time, or were there concurrent operations going or something else? |
Only one - via the compute-redhat.yml :-) I wonder if this would result in the 'kernel panic' eventually. The playbook seems/completes successful but apparently something goes terribly wrong with the image (build?) itself. (side note: I do have to fix rhsm.conf half way through the play within the image itself, otherwise the redhat.repo gets overwritten and redirects to cdn.redhat.com instead - but i doubt that it would result in image issues itself since afterwards everything kicks of fine.) |
w00000t.... i think i may have solved it... What i did was posted here: https://www.linuxquestions.org/questions/linux-server-73/centos-7-does-not-boot-4175619015/ Like so...
I've got no clue whatsoever why it doesn't work without.. but I'll test the compute-redhat.yml again with a new image soon to verify if this actually solved it. |
There were two problems that you hit. There was indeed a bug in the cli where a returned call caused the python trace. That has been solved and will be released soon. The other problem you see, the missing of /sbin/init is something i cannot really explain yet. May i ask when you cloned the TrinityX repo? this helps us determining if this is an ongoing problem or something that has already been solved through other fixes. |
The repo has been cloned (lucky for me I keep track of things using ARA) on the 24th during the re-deployment of the entire controller on RHEL 8.10; |
I too have this issue. I note whilst executing "ansible-playbook default.yml" the following [WARNING]: Target is a chroot or systemd is offline. This can lead to false positives or prevent the init system tools from working. Suspect it might be a systemd not behaving in chroot related issue..... |
A bit more information. I pulled the repro yesterday and built a new controller per the instructions here (https://supercomputing.tue.nl/documentation/administration/trinityx/installation/#fix-uchiwa-logrotate-script-owner ) it completed without issue. I then went to build the node images #ansible-playbook compute-default.yml -v it also completed without error except for the warning above. I have since changed to an alternate distribution build I get the same error regardless. I also note that during the bring up of the node it fails to copy /etc/passwd and /etc/groups |
so I commented out the following section
and the image builds and deploys without error....time to step through the setup and see what breaks |
hi gents, quite a bit has changed (read: fixes, improvements and of course the addition of new bugs :) --Antoine |
Problem description
During iPXE boot, the following challenge pops up. Maybe someone has encountered this before in the past?
Command used
ansible-playbook compute-redhat.yml -v
Expected results
A working image that boots successfully :-)
The text was updated successfully, but these errors were encountered: