Closed Bug 1326237 Opened 7 years ago Closed 7 years ago

Automation Error: mozprocess timed out after 300 seconds running ['python', 'firefox_ui_harness/cli_update.py' ..]

Categories

(Testing :: Firefox UI Tests, defect)

Version 3
All
macOS
defect
Not set
normal

Tracking

(firefox53 affected, firefox54 affected, firefox55 affected)

RESOLVED WORKSFORME
Tracking Status
firefox53 --- affected
firefox54 --- affected
firefox55 --- affected

People

(Reporter: intermittent-bug-filer, Unassigned)

References

Details

(Keywords: intermittent-failure, regression)

Checking the log I can see a coding exception for the shield add-on. Not sure if that is related to this hang but it happens right before:

https://treeherder.mozilla.org/logviewer.html#?job_id=65088966&repo=mozilla-central#L4984

 05:01:10     INFO -  A coding exception was thrown and uncaught in a Task.
 05:01:10     INFO -  Full message: ReferenceError: fetch is not defined
 05:01:10     INFO -  Full stack: apiCall@resource://shield-recipe-client/lib/NormandyApi.jsm:37:5
 05:01:10     INFO -  get@resource://shield-recipe-client/lib/NormandyApi.jsm:44:12
 05:01:10     INFO -  this.NormandyApi.fetchRecipes<@resource://shield-recipe-client/lib/NormandyApi.jsm:52:34
 05:01:10     INFO -  TaskImpl_run@resource://gre/modules/Task.jsm:319:42
05:01:10 INFO - TaskImpl@resource://gre/modules/Task.jsm:277:3

Beside that there is also a network error:

 05:06:41     INFO -  Full message: TypeError: NetworkError when attempting to fetch resource.
05:06:41 INFO - Full stack:

Sadly no stack is shown. I will have to wait for further test results to see what's actually causing this problem.
Summary: Intermittent Automation Error: mozprocess timed out after 300 seconds running ['python', 'firefox_ui_harness/cli_update.py' ..] → Automation Error: mozprocess timed out after 300 seconds running ['python', 'firefox_ui_harness/cli_update.py' ..]
Osmose, can you take a look here?
Flags: needinfo?(mkelly)
Todays update tests show only one or two failures of this type on central and aurora. So the failures massively dropped compared to the last days.
We synced our Github repo to mozilla-central about a week ago (bug 1349348), and it got uplifted to Aurora yesterday. If we were only seeing the issue on Aurora and not on central for the past week then it'd make sense that that sync fixed it. That it was failing on central until today seems to rule that out.

The issue in comment 1 looks like bug 1326225, which was fixed in central in January, so the continued failures probably aren't related to that either.

Without a stack trace (or something in the logs saying that it was a SHIELD test that's failing) that error is hard for me to debug. We've had some issues with correctly mocking out network calls during our tests, but AFAIK those have all been resolved. 

With the failure rate dropped, should we wait to see if this has been resolved already?
Flags: needinfo?(mkelly)
Those are no shield tests, but update tests as run via the firefox-ui-harness which itself is based on Marionette. Those tests do not exercise anything from Shield/Normandie. The failures were just something I noticed when checking the log files for the tests around the time when the failure happened. So they might not be related at all, but who knows.

I will continue to observe the failures as usual. If they stick with this lower failure rate I might find some time next week to maybe check it more closely.
Until bug 1355009 is not fixed, this failure might be hidden.
Depends on: 1355009
Checking Orange Factor this failure is really gone from Aurora. On that branch we were not affected by bug 1355009. So I believe that this is fixed now.

https://brasstacks.mozilla.com/orangefactor/?display=Bug&bugid=1326237&startday=2017-03-15&endday=2017-04-12&tree=all

Marking as WFM given that we don't actually know what fixed it.
Status: NEW → RESOLVED
Closed: 7 years ago
Resolution: --- → WORKSFORME
You need to log in before you can comment on or make changes to this bug.