My test job .yaml does not provide any error while health-check trigger
from GUI and give error while submitting the same test job from the command
you can find the command line submission error as mentioned below.
lavamaster@lava$lavacli -i admin@validation jobs submit /tmp/test.yaml
Unable to submit /tmp/target.yaml: <Fault 400: "Problem with submitted job
data: expected str for dictionary value @
I have reported same type of bug for "soft-reboot" command support and
provided the patch by lava-team for this soft-reboot support issue i.e "
Can someone provide the path for "shutdown-message" support also?
Hello, I have a question about "count" in lava multinode, for next:
What does "count" mean? I know the job will only be scheduled if there are 2 "xxx" based device idle.
But, after that, how I could use one device do one task, and use another device to do another task? How can I assign different tasks to the 2 devices with the same devicetype?
device xxx1 based on xxx
device xxx2 based on xxx
Then xxx1 run "ls" while xxx2 run "uname"?
We upgrade lava 2018 to 2021, we met the problem when we submit jobs：
Job error: Invalid job data: ['Device not configured to support serial connection.', 'Device not configured to support serial connection.']
Hello lava users,
I have to deal with a board with a very limited storage size. Overall capacity is 64 MB, with about 30 MB remaining.
This is a clear limitation for Lava usage. Hence my questions:
* Is there a possibility to configure Lava tests installation on an external storage, such as an USB drive?
* In Lava recent versions, is it possible to optimize the overlay tests structure?
* Today, assuming that tests are stored in a git repo, for each test the complete repo is copied in each overlay
* A balance has to be made between one git containing all the tests and one git per test (extreme configurations)
* Is there a better way to manage this? In a ideal world, that would mean copy only what is needed and not more
Thanks, have a nice day
I'm hosting a lava-setup for us which is based upon the lava-docker-compose setup from here https://git.lavasoftware.org/lava/pkg/docker-compose
I managed to configure Django to use a EMAIL_HOST which allows me to notify my users. However dbutils.get_domain() returns example.com and some digging lead me to the site setup from Django. I'm not entirely sure how it's intended to work but I would expect that LAVA fetches the host to use from .env, where I specified DC_LAVA_SERVER_HOSTNAME=lava.mydomain.com.
Where would I set that or what did I miss?
I figured out a few details about the docker-compose setup by now, eg how to configure LDAP. Where would I add such documentation? A PR to the docs?
Best regards, Olli
While testing 2020.10 with https enabled, we have noticed the following problem:
In /etc/lava-dispatcher/lava-worker, if we set WS_URL="--ws-url http://<master's address>/ws/" (or do not this variable at all), we get the following error on worker:
2020-11-19 15:14:57,751 INFO PING => server
SSL error in data received
protocol: <asyncio.sslproto.SSLProtocol object at 0x7f9a0d704160>
transport: <_SelectorSocketTransport fd=8 read=polling write=<idle, bufsize=0>>
Traceback (most recent call last):
File "/usr/lib/python3.7/asyncio/sslproto.py", line 526, in data_received
ssldata, appdata = self._sslpipe.feed_ssldata(data)
File "/usr/lib/python3.7/asyncio/sslproto.py", line 207, in feed_ssldata
File "/usr/lib/python3.7/ssl.py", line 767, in unwrap
ssl.SSLError: [SSL: KRB5_S_INIT] application data after close notify (_ssl.c:2609)
Still, even with those errors the worker is fully functional.
Now, the only way to get rid of such errors is to explicitly specify the port, 8001:
WS_URL="--ws-url http://<master's address>:8001/ws/"
Have you experienced similar problems? Did we miss something on our setup?
Thanks a lot,