TASKCLUSTER_ROOT_URL is not set on terraform-packet
Categories
(Firefox Build System :: Task Configuration, defect)
Tracking
(firefox71 fixed)
Tracking | Status | |
---|---|---|
firefox71 | --- | fixed |
People
(Reporter: glandium, Assigned: glandium)
References
Details
Attachments
(1 file)
I got failures on Android x86 tests that run on packet.net when I added fetches:
https://treeherder.mozilla.org/logviewer.html#/jobs?job_id=239839533&repo=try&lineNumber=31
[taskcluster 2019-04-12 03:49:44.837Z] Worker ID: machine-8
[taskcluster 2019-04-12 03:49:44.837Z] Worker Group: packet-sjc1
[taskcluster 2019-04-12 03:49:44.837Z] Worker Node Type: packet.net
[taskcluster 2019-04-12 03:49:44.837Z] Worker Type: gecko-t-linux
[taskcluster 2019-04-12 03:49:44.837Z] Public IP: 147.75.109.15
[taskcluster 2019-04-12 03:49:44.837Z] using cache "gecko-level-1-checkouts-v3-33ea6ead87f10b63cd64" -> /builds/worker/checkouts
[taskcluster 2019-04-12 03:49:44.837Z] using cache "gecko-level-1-tooltool-cache-v3-33ea6ead87f10b63cd64" -> /builds/worker/tooltool-cache
[taskcluster 2019-04-12 03:49:46.610Z] Image 'public/image.tar.zst' from task 'XixQ3m-oS3C-95iK9P4Gmg' loaded. Using image ID sha256:6a0e1b6ec6889142f31cb0a5a8b1a208a3e7ed2d133eb89372126fb89b97761f.
[taskcluster 2019-04-12 03:49:46.974Z] === Task Starting ===
[setup 2019-04-12T03:49:47.440Z] run-task started in /builds/worker
[cache 2019-04-12T03:49:47.442Z] cache /builds/worker/checkouts exists; requirements: gid=1000 uid=1000 version=1
[cache 2019-04-12T03:49:47.442Z] cache /builds/worker/tooltool-cache exists; requirements: gid=1000 uid=1000 version=1
[volume 2019-04-12T03:49:47.442Z] changing ownership of volume /builds/worker/.cache to 1000:1000
[volume 2019-04-12T03:49:47.442Z] volume /builds/worker/checkouts is a cache
[volume 2019-04-12T03:49:47.442Z] volume /builds/worker/tooltool-cache is a cache
[volume 2019-04-12T03:49:47.442Z] changing ownership of volume /builds/worker/workspace to 1000:1000
[setup 2019-04-12T03:49:47.442Z] running as worker:worker
[fetches 2019-04-12T03:49:47.442Z] fetching artifacts
[fetches 2019-04-12T03:49:47.442Z] executing ['/usr/bin/python3', '-u', '/builds/worker/bin/fetch-content', 'task-artifacts']
Traceback (most recent call last):
File "/builds/worker/bin/fetch-content", line 443, in <module>
sys.exit(main())
File "/builds/worker/bin/fetch-content", line 439, in main
return args.func(args)
File "/builds/worker/bin/fetch-content", line 393, in command_task_artifacts
root_url = os.environ['TASKCLUSTER_ROOT_URL']
File "/usr/lib/python3.5/os.py", line 725, in __getitem__
raise KeyError(key) from None
KeyError: 'TASKCLUSTER_ROOT_URL'
Comment 1•6 years ago
|
||
I'm guessing that they're running a fairly old version of whatever worker implementation they use. As a workaround, you can set TASKCLUSTER_ROOT_URL in task.payload.env
, but the better fix will be an upgrade of the worker.
Updated•6 years ago
|
Assignee | ||
Updated•6 years ago
|
Assignee | ||
Comment 2•6 years ago
|
||
Comment 3•6 years ago
|
||
Any chance we could get these workers updated to use a new enough (or patched) worker that sets this?
Comment 4•6 years ago
|
||
Looking at https://github.com/taskcluster/taskcluster-infrastructure/blob/master/modules/docker-worker/main.tf#L131, I see TASKCLUSTER_ROOT_URL being set, but I confess I don't know whether that's actually propagated into the worker.
Wander: what do we need to do to make this right?
Comment 5•6 years ago
|
||
The required change landed in https://github.com/taskcluster/docker-worker/pull/418, if that helps.
Comment 6•6 years ago
|
||
(In reply to Chris Cooper [:coop] pronoun: he from comment #4)
Looking at https://github.com/taskcluster/taskcluster-infrastructure/blob/master/modules/docker-worker/main.tf#L131, I see TASKCLUSTER_ROOT_URL being set, but I confess I don't know whether that's actually propagated into the worker.
Wander: what do we need to do to make this right?
it is fixed in custom images. Once it lands, we are good.
Comment 7•6 years ago
|
||
:wcosta Do you have an ETA for this landing? It is blocking other worker.
Comment 8•6 years ago
|
||
(In reply to Tom Prince [:tomprince] from comment #7)
:wcosta Do you have an ETA for this landing? It is blocking other worker.
I am close to getting new images done, but I am busy right now with GCP migration and ops hand off.
Comment 9•6 years ago
|
||
There's a r+ patch which didn't land and no activity in this bug for 2 weeks.
:glandium, could you have a look please?
For more information, please visit auto_nag documentation.
Assignee | ||
Comment 10•6 years ago
|
||
Should I just land the workaround patch, then?
Comment 11•6 years ago
|
||
(In reply to Mike Hommey [:glandium] from comment #10)
Should I just land the workaround patch, then?
Yes, please. We are currently evaluating AWS bare metal as a replacement for packet.net.
Comment 12•6 years ago
|
||
Comment 13•6 years ago
|
||
bugherder |
Description
•