Hi,
I saw something like "LAVA_HARD_RESET_COMMAND" was added to environment.
Yes, the test action could use this command to reboot the device now, it's useful.
But, after the command sent, the code in container still need to handle something like next:
"Expect press any key to continue"
"Press any key"
"Expect login prompt"
"Login"
Etc......
But, this exactly the same code what lava have do in a lots of python files located in "lava_dispatcher/actions/boot".
My question is: do you think it's possible for user code in "docker-test-shell" action directly send any command or use any protocol to tell lava code(maybe a listener or something else) to help to do these things, then after done, the user code in docker test shell continue to execute its own stuff?
You know, this could reuse existed lava code as much as possible. I just want to know if you have plan for this? Thanks.
Hi,
I have some questions about the fastboot/adb in docker support I hope you can help with.
Use case is for android 10 aosp testing with LAVA 2020.05.
Thanks to Antonio's presentation and draft documentation I have simple fastboot host to
DUT communication working for a u-boot based arm64 board. I am now trying to apply an
existing flash process, which uses a script on the host to send fastboot cmds, into a LAVA job.
I can see how fastboot --set-active, reboot and flash commands all have equivalent controls
in a 'deploy to fastboot docker' LAVA job section. Do equivalents exist for the fastboot
oem format, oem erase, format and erase commands or is there a way to insert them
in the deploy?
Expecting that will take some engineering work, in parallel I wanted as a stop gap to try
running the flash script from a LAVA job. So people could work on the testing side whilst
I resolved the deploy section. Antonio suggested trying to do that from the test section
I recall.
To do that I face two issues:
1) The build artifacts are on a local volume rather than an artifact server so I need to
get them into the docker container in an automated way. Is there a way to either mount
a local volume or file list into the container without asking LAVA to deploy them?
As an experiment I tried using the docker image manipulation added in 2020.04 to do
this. There I hit a problem with the current implementation. It seems the 'deploy to
downloads' implementation does not check for a local image first, as the other docker
support does, before trying to pull the image. So I get an error when the pull fails:
https://lava.genivi.org/scheduler/job/961#L24
2) The job needs to be able to boot the DUT, interrupt u-boot and start fastboot
so the host has something to communicate with once the test section is reached.
I can achieve that in a horrible hacky way by having a deploy section with a simple
single flash image (dtbo say) and using the reboot controls to get the board to reboot into
fastboot to await the flash script being run from the test section, but I expect there
is a better way. Any ideas?
Regards
Steve
Hi,
Taking the current master branch head (0f6e89e3) I find my health checks
that use u-boot tftboot support are now failing with the following:
start: 2.2 bootloader-overlay (timeout 00:05:00) [common]
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/lava_dispatcher/action.py", line 245, in run_actions
new_connection = action.run(connection, action_max_end_time)
File "/usr/lib/python3/dist-packages/lava_dispatcher/actions/boot/__init__.py", line 454, in run
tee_addr = self.job.device["parameters"][self.bootcommand]["tee"]
KeyError: 'tee'
Example job:
https://lava.genivi.org/scheduler/job/1160
I've not combed through the code in detail but it looks like it is related
to the recent tee merge:
https://git.lavasoftware.org/lava/lava/-/commit/f18ca091e61da31159cd72a0030…
That appeared to introduce an optional feature which my jobs do not
make use. Am I missing something?
Regards
Steve
Dear lava-users,
I'm intern at /e/ foundation and working on using LAVA to test our OS updates. We are trying to configure our system in order to use the recent feature allowing the use of docker instead of LXC containers. On this point we are grateful to the developpers for adding this possibility because we were having problems with LXC.
Now we are having issues because we can't flash our OS through fastboot, we have to use adb sideload with .zip archives. By the way, we can't do it through the classic deploy section. We came with the idea of just flashing a twrp recovery partition via the deploy section using fastboot, and then trying to run shell commands in the test section to be able to use adb sideload and install our ROM.
For the moment as you can see in the job we didn't add all the shell commands because we already have some problems with this simple try. The objective would be to pass this booting into recovery step and then try to continue with adb sideload. I've also joined the logs and the device and device type dictionaries. I would like to have your opinion and advices on what we are trying to achieve.
Best regards,
Baptiste and the /e/ team
Hello everyone,
I would like to integrate the u-boot tests suite (pytest) in lava. ( https://github.com/u-boot/u-boot/tree/master/test/py)
We are now executing the test suite manually, using a host machine and the DUT.
The u-boot tests suite is installed on a host computer and the DUT is "stopped" in uboot environment.
The host computer execute the "u-boot pytest" framework, commands are sent to the DUT and tests results are collected to the host computer.
It is a "semi automatic" tests and I would like to do a full automation using LAVA.
I plan to use multinode jobs, one job instances for the DUT and another one for the remote host computer.
So, I would like to know if such tests have already been craeted in LAVA, and if someone could give me advice on this tests implementation.
My Lava version is: 2020.06
Best regards
Philippe Begnic
STMicroelectronics
Hello,
We are wanting to update to a more recent LAVA docker release, we've been at 2019.12 since it was released. Our configuration relies on lxc, and after moving to 2020.01 all of our test jobs hit the error "lxc container 'lxc-test-17071' already exists" very early on (17071 is the test job number). I've pasted our device configuration, test job, and log output in the below pastebin. What should we do to keep our LAVA jobs working with the 2020.01 release and beyond?
https://pastebin.com/UQixc6RS
Thanks,
Alex
This message and any attachments may contain confidential information from Cypress or its subsidiaries. If it has been received in error, please advise the sender and immediately delete this message.