Closed
Bug 1204281
Opened 9 years ago
Closed 7 years ago
Intermittent [taskcluster:error] Task timeout after 3600 seconds. Force killing container. / [taskcluster:error] Task timeout after 5400 seconds. Force killing container. / [taskcluster:error] Task timeout after 7200 seconds. Force killing container.
Categories
(Release Engineering :: General, defect)
Release Engineering
General
Tracking
(firefox51 fixed)
RESOLVED
DUPLICATE
of bug 1411358
Tracking | Status | |
---|---|---|
firefox51 | --- | fixed |
People
(Reporter: philor, Assigned: gbrown)
References
(Depends on 3 open bugs)
Details
(Keywords: intermittent-failure, Whiteboard: [stockwell infra])
Attachments
(5 files, 1 obsolete file)
(deleted),
text/x-review-board-request
|
dminor
:
review+
|
Details |
(deleted),
patch
|
dminor
:
review+
|
Details | Diff | Splinter Review |
(deleted),
patch
|
dminor
:
review+
|
Details | Diff | Splinter Review |
(deleted),
patch
|
cbook
:
review+
|
Details | Diff | Splinter Review |
(deleted),
patch
|
jmaher
:
review+
|
Details | Diff | Splinter Review |
+++ This bug was initially created as a clone of Bug #1198092 +++
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Legacy TBPL/Treeherder Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 53•9 years ago
|
||
I think this has to do with Gij's retry logic. As Michael said on the mailing list, each Gij test is run 5 times within a test chunk (g. Gij4) before it is marked as failing. Then that chunk itself is retried up to 5 times before the whole thing is marked as failing.[1] This means we may end up retrying so much that we timeout the task.
I wouldn't consider this a bug for TC but rather a bug for b2g Gij automation. What do you guys think about us changing the component here?
[1]: https://groups.google.com/d/msg/mozilla.dev.fxos/LTTobhx4tCc/nN_gad51AgAJ
Flags: needinfo?(mhenretty)
Comment 54•9 years ago
|
||
(In reply to Nigel Babu [:nigelb] from comment #53)
> I think this has to do with Gij's retry logic. As Michael said on the
> mailing list, each Gij test is run 5 times within a test chunk (g. Gij4)
> before it is marked as failing. Then that chunk itself is retried up to 5
> times before the whole thing is marked as failing.[1] This means we may end
> up retrying so much that we timeout the task.
>
> I wouldn't consider this a bug for TC but rather a bug for b2g Gij
> automation. What do you guys think about us changing the component here?
>
> [1]:
> https://groups.google.com/d/msg/mozilla.dev.fxos/LTTobhx4tCc/nN_gad51AgAJ
Certainly our retry logic makes this worse, but the real problem is that for these bad (ie. long) runs, something happens which makes the test runner pause 11 minutes between test runs. If you take a look at one of the failures [1], there was really only 1 test that was being retried, apps/system/test/marionette/text_selection_test.js. But several tests before that (which passed the first time) would have an 11 minute delay between tests. This is where the problem is.
Now this could still totally be a bug in our Gij automation rather than taskcluster, but we should investigate what is happening during those 11 minutes before drawing any conclusions.
1.) https://public-artifacts.taskcluster.net/XMYzs9RrQyyVTjlQy7d_Aw/0/public/logs/live_backing.log
Flags: needinfo?(mhenretty)
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 64•9 years ago
|
||
:rail it looks like some of the spikes recently were related to funsize tasks timing out, such as https://tools.taskcluster.net/task-inspector/#WXTTC2IoQ-WesLVclUDX7A/0 . Most oft he tasks seem to fail with "HTTPError: 400 Client Error: BAD REQUEST"
The last week or so of tasks, there were 755 failures and 625 of them were from a funsize task. I wasn't sure if you were aware of anything going on. Let me know if there is anything I can dig into with this.
Flags: needinfo?(rail)
Comment 65•9 years ago
|
||
I filed bug 1223872 to make balrog submission have less race conditions, but I'm not sure if it can be quickly and easily resolved.
There is also bug 1224698 to help with networking issues we have after aus migrated to scl3, but still routed through phx.
Flags: needinfo?(rail)
Comment 66•9 years ago
|
||
Thanks for the update! At least there are bugs out there to fix this. I just didn't know if you were aware of the spike or not. Thanks again!
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 85•9 years ago
|
||
looking at the last few items here, I see we timeout when the entire job is >60 minutes- not the tests. This includes setup, running tests, cleanup.
I see 27 minutes for setup due to vcs errors:
https://treeherder.mozilla.org/logviewer.html#?repo=fx-team&job_id=6263182
https://treeherder.mozilla.org/logviewer.html#?repo=fx-team&job_id=6263184
https://treeherder.mozilla.org/logviewer.html#?repo=fx-team&job_id=6263184
all of these started at roughly the same time, so this looks to be a root issue in the time to complete:
curl --connect-timeout 30 --speed-limit 500000 -L -o /home/worker/.tc-vcs/clones/hg.mozilla.org/integration/gaia-central.tar.gz https://queue.taskcluster.net/v1/task/Dq4miM-9T6ygRTP9h-XWkQ/artifacts/public/hg.mozilla.org/integration/gaia-central.tar.gz
this should normally take <5 minutes and it shows 20+ minutes in many logs.
:rail, can you help me find the person who would know why the vcs sync is so slow- I assume this was an isolated incident.
Flags: needinfo?(rail)
Comment 86•9 years ago
|
||
then I see a few others where we have no vcs issues and I see:
14:35:39 INFO - Running tests in /home/worker/gaia/apps/system/test/marionette/audio_channel_competing_test.js
15:03:25 INFO - .....................................................................................................................................................
15:03:25 INFO - /home/worker/gaia/apps/system/test/marionette/audio_channel_competing_test.js failed. Will retry.
15:03:30 INFO - Running tests in /home/worker/gaia/apps/system/test/marionette/audio_channel_competing_test.js
[taskcluster:error] Task timeout after 3600 seconds. Force killing container.
this seems as though that specific test case is failing.
:bkelly, can you help me find the owner of gaia/apps/system/test/marionette/audio_channel_competing_test.js, I cannot find it in dxr, and this must live somewhere in b2g land.
Flags: needinfo?(bkelly)
Comment 87•9 years ago
|
||
(In reply to Joel Maher (:jmaher) from comment #85)
> :rail, can you help me find the person who would know why the vcs sync is so
> slow- I assume this was an isolated incident.
I'd talk to hwine.
Flags: needinfo?(rail)
Comment 88•9 years ago
|
||
(In reply to Joel Maher (:jmaher) from comment #86)
> :bkelly, can you help me find the owner of
> gaia/apps/system/test/marionette/audio_channel_competing_test.js, I cannot
> find it in dxr, and this must live somewhere in b2g land.
I emailed gregor and mhenretty. I think those are bug 1233565. Thanks for checking on this!
Flags: needinfo?(bkelly)
Comment 89•9 years ago
|
||
Right, we do have an owner for gaia/apps/system/test/marionette/audio_channel_competing_test.js in bug 1233565. But also note that this test has been disabled for about a week [1], and we have still been seeing this failure in automation since. So I still think this issue is caused by some large slowdown in the test runner (maybe the VM gets choked in amazon or something), and not due to any individual test(s).
1.) https://github.com/mozilla-b2g/gaia/commit/0ffdc828b44dc33b84a3b34ce1643102d04b116a
Comment 90•9 years ago
|
||
:hwine, when you get back, could you look at why this vcs sync is taking so long intermittently? In fact, I would rather kill the job if we have a vcs sync error that takes so long as it would terminate the job faster since we know it will always terminate. Can we define a guarantee of 6 minutes for all source syncing and if we cross that threshold we fail?
Flags: needinfo?(hwine)
Comment hidden (Intermittent Failures Robot) |
Comment 92•9 years ago
|
||
(In reply to Joel Maher (:jmaher) from comment #90)
> :hwine, when you get back, could you look at why this vcs sync is taking so
> long intermittently?
Can you define "this", please? There are many (>500) vcs-sync jobs. There is lots of tuning that can be done once we know which repos are involved.
> In fact, I would rather kill the job if we have a vcs
> sync error that takes so long as it would terminate the job faster since we
> know it will always terminate. Can we define a guarantee of 6 minutes for
> all source syncing and if we cross that threshold we fail?
I'm not sure I'm following here - there should be no time based dependencies in vcs-sync (it is supposed to be event driven). So "all source syncing" doesn't mean anything to me. Let's do a vidyo to educate me.
However, no guarantee of that short (6m) a time. See https://wiki.mozilla.org/User:Hwine/Holiday_VCS-Sync_Troubleshooting#Diagnosing_Single_Repo_Issues for a diagram of what is happening, and how there are some events that should gate potential race conditions.
Flags: needinfo?(hwine) → needinfo?(jmaher)
Comment 93•9 years ago
|
||
this is more of an issue with curling a repo from https://queue.taskcluster.net:
https://public-artifacts.taskcluster.net/KoAcWC4BR62evcQkVSXQ1A/0/public/logs/live_backing.log (search for "Operation too slow")
as per comment 85, this seems to be clustered at certain times. The link at the top of this current comment is from 2 days ago and was the only instance. Is this something that has a capacity of XX connections/second and we hit a perfect storm every now and then?
:garndt, how can we debug this and ensure that the .gz is available and we can determine where the proper caches are updated/out of date? Maybe :gps would know more details.
Flags: needinfo?(jmaher) → needinfo?(garndt)
Comment 94•9 years ago
|
||
I have setup monitoring of some of our taskcluster-vcs cached repos (basically everything but emulator and device image 'repo' repos, those will come soon with a patch) so we should notice hopefully when one of those are out of date (> 48 hours old). Caches expire after 30 days so for a new cache task to not be indexed in 48 hours is enough time to look into it as long as we receive the alert.
As far as ensuring that the .gz is available, it is in this case because it started the transfer. Also, in tc-vcs >2.3.17 it will give an error if the artifact couldn't be found for a particular indexed task. I'm in the process of upgrading our builders and phone builder images with that, and then can move onto the tester image used here.
In this case, the artifact fact was found and was being downloaded, but it was just too slow (<500 kB/s) for too long (> 30 seconds I think).
Looking at when this task was run and when the cache for gaia was created, it's possible that there was slow transfer from us-west-1 (where the ec2 instance was) and us-west-2 (where the artifact in s3 lives). If I recall correctly (I would need to double check this), our s3-copy-proxy will copy to the requesting region but in the meantime, will redirect the client to the canonical region in us-west-2 for an artifact until the copy is completed.
Flags: needinfo?(garndt)
Comment 95•9 years ago
|
||
nice, it sounds like we have a proactive solution to this. Now to figure out how to get test timeouts in a different bug :) I will wait to see how many issues show up with this bug over the next couple of days.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 100•9 years ago
|
||
:garndt, we are seeing a lot of these errors still, can you check your monitoring and see if anything stands out? Maybe we have small time windows where a lot happens, 16 issues today, 246 yesterday (Sunday), 24 last Friday.
The more issues we can pinpoint the better!
Flags: needinfo?(garndt)
Comment 101•9 years ago
|
||
Spot checking some of these it appears a majority of the time is spent within the tests (usually around 56-58 minutes of the 1 hour max run time) and eventually time out.
I checked out some metrics related to these instances, and it appears that a majority of them are running out of memory.
Is the spike in these failing around the same time a majority of these instances on m1.medium?
Flags: needinfo?(garndt)
Comment 102•9 years ago
|
||
ah, I overlooked the obvious! lets split the R(J) jobs into 2 chunks, :armenzg, can you look at splitting Jsreftests into 2 chunks?
Flags: needinfo?(armenzg)
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 106•9 years ago
|
||
Armen, can we do mochitest-plain in more chunks, we are hitting the 3600 second timeout on chunk 4 many times. Maybe 15? 5 of the jobs are >40 minutes, most in the 50+ minute range with chunk 4 at 55 minutes+. The other 5 are in the 15-22 minute range. We should verify that --chunk-by-runtime is defined as well.
Flags: needinfo?(armenzg)
Comment hidden (Intermittent Failures Robot) |
Comment 109•9 years ago
|
||
A change landed yesterday and it got merged today.
We will have few more instances but this should quite down (including today):
https://brasstacks.mozilla.com/orangefactor/?display=Bug&bugid=1204281&startday=2016-01-25&endday=2016-01-26&tree=all
Comment hidden (Intermittent Failures Robot) |
Comment 111•9 years ago
|
||
If we ignore the first day of the range we go down to 16:
https://brasstacks.mozilla.com/orangefactor/?display=Bug&bugid=1204281&startday=2016-01-26&endday=2016-01-31&tree=all
If we ignore one more day we're down to 6:
https://brasstacks.mozilla.com/orangefactor/?display=Bug&bugid=1204281&startday=2016-01-27&endday=2016-01-31&tree=all
As we enable more jobs we will keep an eye on this.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 114•9 years ago
|
||
rail, funzise tasks have started spiking up (scroll down):
[funsize] Publish to Balrog (today-2, chunk 4, subchunk 1)
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 118•9 years ago
|
||
It seems that funsize accounts for more than half of all these occurrences and most occurred on the 26th.
Flags: needinfo?(rail)
Comment 119•9 years ago
|
||
I looked at those and most of them are timeout due to balrog submission retries. This is a known issue and aki is going to look at new worker type for balrog submission.
Flags: needinfo?(rail)
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 122•9 years ago
|
||
wow, 75 failures in the last week on Aurora! This looks all related to funsize stuff. :bhearsum, can you take a look at this?
Flags: needinfo?(bhearsum)
Comment 123•9 years ago
|
||
(In reply to Joel Maher (:jmaher) from comment #122)
> wow, 75 failures in the last week on Aurora! This looks all related to
> funsize stuff. :bhearsum, can you take a look at this?
It looks like Aki is going to be looking at this (maybe in a roundabout way) soon, based on comment #119.
Flags: needinfo?(bhearsum) → needinfo?(rail)
Comment 124•9 years ago
|
||
We can retry harder, but I'm not sure if it's going to be better...
Flags: needinfo?(rail)
Comment 125•9 years ago
|
||
ok, if :aki is reworking the balrog submission that sounds like it should resolve this funsize stuff. Is this a March thing or a Q2 thing? This specific error is pretty high on the orange factor list.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 128•9 years ago
|
||
:aki, can you comment on when your work for the new balrog worker will be completed and in pace?
:rail, is there anything else we can do in the meantime?
Flags: needinfo?(aki)
Comment 129•9 years ago
|
||
:jmaher, currently this looks like a [mid-?]q2 thing.
Flags: needinfo?(aki)
Comment 130•9 years ago
|
||
thanks :aki! I wonder if there are ways to reduce this error outside of the new balrog worker in the short term. If not, this doesn't seem to be getting much worse, it is still one of the top issues the sheriffs have to star though.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Reporter | ||
Updated•9 years ago
|
Summary: Intermittent [taskcluster:error] Task timeout after 3600 seconds. Force killing container. → Intermittent [taskcluster:error] Task timeout after 3600 seconds. Force killing container. / [taskcluster:error] Task timeout after 5400 seconds. Force killing container.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 139•9 years ago
|
||
This got very frequent. Joel, can you investigate if this caused by a performance regression, please?
Flags: needinfo?(jmaher)
Comment 140•9 years ago
|
||
ok this is 2 issues:
* linux64 debug xpcshell (chunk 4/5 are hitting the 1 hour limit)- this is already fixed as this is now 10 chunks
* linux64 debug mda, 1 chunk, hitting the 55+ minutes normally
I am testing a patch on try to see if I can split mda into 2 chunks
Comment hidden (Intermittent Failures Robot) |
Comment 142•9 years ago
|
||
ok, mda needs 3 chunks but doing so yields a failure in test_zmedia_cleanup.html:
https://treeherder.mozilla.org/#/jobs?repo=try&revision=081c2ddc2501a54467d3e86be0a91ee837ac2bf5
I am thinking it might be better to just extend the timeout.
:dminor, can you weigh in here and figure out if we need a longer timeout or should split this up and make it green?
Flags: needinfo?(jmaher) → needinfo?(dminor)
Comment 143•9 years ago
|
||
So it looks like test_zmedia_cleanup.html was added as a hacky of way of cleaning up network state for B2G testing and doesn't actually test anything. We have a Bug 1188120 on file to remove it, I'm going to see if we can go ahead and do that. I've hit lots of intermittent failures with that test.
Flags: needinfo?(dminor)
Comment 144•9 years ago
|
||
:dminor, would you be fine splitting this to 3 chunks for linux64 debug?
Comment 145•9 years ago
|
||
dminor, or would you rather see a longer timeout?
Depends on: 1188120
Flags: needinfo?(dminor)
Comment 147•9 years ago
|
||
Review commit: https://reviewboard.mozilla.org/r/53970/diff/#index_header
See other reviews: https://reviewboard.mozilla.org/r/53970/
Attachment #8754427 -
Flags: review?(dminor)
Comment 148•9 years ago
|
||
Comment on attachment 8754427 [details]
MozReview Request: Bug 1204281 - split linux64 debug taskcluster M(mda) into 3 chunks. r?dminor
https://reviewboard.mozilla.org/r/53970/#review50682
Please revise the commit message to reflect changing the timeout rather than splitting the test into three chunks.
Attachment #8754427 -
Flags: review?(dminor) → review+
Updated•9 years ago
|
Keywords: leave-open
Comment 149•9 years ago
|
||
Comment 150•9 years ago
|
||
Backed out for breaking gecko decision task: https://hg.mozilla.org/integration/mozilla-inbound/rev/2b227a22287677ac7af098166a632e768e70d022
Push with failures: https://treeherder.mozilla.org/#/jobs?repo=mozilla-inbound&revision=03ed23408215dbc98f987c68a568af89adb25eb8
Flags: needinfo?(jmaher)
Comment 151•9 years ago
|
||
last patch had an indentation problem and broke the tree! this is what I get for landing code after just removing unused lines from a patch and accidentally hitting the space bar.
also mozreview doesn't work here as there is already some parent review request.
Flags: needinfo?(jmaher)
Attachment #8754532 -
Flags: review?(dminor)
Comment 152•9 years ago
|
||
Comment on attachment 8754532 [details] [diff] [review]
increase timeout to 5400 seconds
Review of attachment 8754532 [details] [diff] [review]:
-----------------------------------------------------------------
lgtm
Attachment #8754532 -
Flags: review?(dminor) → review+
Updated•9 years ago
|
Keywords: checkin-needed
Comment hidden (Intermittent Failures Robot) |
Comment 154•9 years ago
|
||
Keywords: checkin-needed
Comment hidden (Intermittent Failures Robot) |
Comment 156•9 years ago
|
||
:dustin, can you help me out here? I extended the timeout from 3600 -> 5400 and we are still getting 5400 second timeouts.
example job:
https://treeherder.mozilla.org/logviewer.html#?repo=mozilla-inbound&job_id=28622438#L44951
code I used to fix the timeout:
https://hg.mozilla.org/integration/mozilla-inbound/rev/823f49140d69
I tested this on try server with a few different cycles and never saw the timeout, but it could have been luck that my jobs finished in <60 minutes.
Flags: needinfo?(dustin)
Comment 157•9 years ago
|
||
The maxRunTime on https://tools.taskcluster.net/task-inspector/#Nggm4AmrRPOmmw27HgBrbg/ is still 3600. However, timeout, which is not a parameter that means anything to docker-worker, is set to 5400. I think you want to set maxRunTime to 5400 :)
Flags: needinfo?(dustin)
Comment 158•9 years ago
|
||
thanks Dustin for the pointer, this should resolve things.
Attachment #8756500 -
Flags: review?(dminor)
Updated•9 years ago
|
Attachment #8756500 -
Flags: review?(dminor) → review+
Comment 159•9 years ago
|
||
bugherder |
Comment hidden (Intermittent Failures Robot) |
Comment 161•9 years ago
|
||
Comment 162•9 years ago
|
||
bugherder |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 171•8 years ago
|
||
current issue is that we run out of time on the job- it is normally 55 minutes- take some slowdown for a failed test or cleanup/symbols and we cross the 60 minute threshold.
What really makes this difficult is that the times are so variable due to the docker image download/setup. I cannot just load the meta data in treeherder, I have to click on each log file which takes a long time (i.e. this random docker setup is making life harder)
My overall impression here is that we need 1 or 2 more chunks, but for now I would like to just bump this up to 90 minutes. It is foolish to split this into more chunks until we can realistically reduce the 20 minute blocks of randomness in the docker setup.
Comment 172•8 years ago
|
||
Attachment #8763502 -
Flags: review?(cbook)
Comment 173•8 years ago
|
||
Comment on attachment 8763502 [details] [diff] [review]
increase browser-chrome timeout from 60 to 90 minutes
Review of attachment 8763502 [details] [diff] [review]:
-----------------------------------------------------------------
looks good, thanks joel!
Attachment #8763502 -
Flags: review?(cbook) → review+
Comment 174•8 years ago
|
||
Pushed by jmaher@mozilla.com:
https://hg.mozilla.org/integration/mozilla-inbound/rev/f89175185de0
90 minute timeout for linux64 mochitest-browser-chrome chunks. r=Tomcat
Comment 175•8 years ago
|
||
Backed out for gecko-decision opt failures:
https://hg.mozilla.org/integration/mozilla-inbound/rev/bc341233192c
Push with failure: https://treeherder.mozilla.org/#/jobs?repo=mozilla-inbound&revision=f89175185de0a20650609a24ead315dc02c4e01f
Traceback (most recent call last):
File "/workspace/gecko/taskcluster/mach_commands.py", line 151, in taskgraph_decision
return taskgraph.decision.taskgraph_decision(options)
File "/workspace/gecko/taskcluster/taskgraph/decision.py", line 79, in taskgraph_decision
create_tasks(tgg.optimized_task_graph, tgg.label_to_taskid)
File "/workspace/gecko/taskcluster/taskgraph/create.py", line 61, in create_tasks
f.result()
File "/workspace/gecko/python/futures/concurrent/futures/_base.py", line 396, in result
return self.__get_result()
File "/workspace/gecko/python/futures/concurrent/futures/thread.py", line 55, in run
result = self.fn(*self.args, **self.kwargs)
File "/workspace/gecko/taskcluster/taskgraph/create.py", line 73, in _create_task
res.raise_for_status()
File "/workspace/gecko/python/requests/requests/models.py", line 840, in raise_for_status
raise HTTPError(http_error_msg, response=self)
HTTPError: 400 Client Error: Bad Request for url: http://taskcluster/queue/v1/task/UINhsKfoSF-0VKRdHS6uVA
Flags: needinfo?(jmaher)
Comment 176•8 years ago
|
||
Hi Dustin, can you take a look at the patch and check what's wrong with it? Thank you.
Flags: needinfo?(dustin)
Comment 177•8 years ago
|
||
I believe the problem with the patch is that I had 2 space indentation vs 4 space indentation (wrong scope...call it scope creep!).
Flags: needinfo?(jmaher)
Comment 178•8 years ago
|
||
The actual error was quite a ways up in the logfile, and indicated a JSON schema failure because maxRunTime was added at the task, rather than task.payload, level.
Flags: needinfo?(dustin)
Comment 179•8 years ago
|
||
ok, this has proper spacing and green on try:
https://treeherder.mozilla.org/#/jobs?repo=try&revision=703cec16666126931b77b69eef4ccfbc09b7b237
Attachment #8763502 -
Attachment is obsolete: true
Attachment #8763826 -
Flags: review?(cbook)
Comment 180•8 years ago
|
||
Comment on attachment 8763826 [details] [diff] [review]
increase browser-chrome timeout from 60 to 90 minutes (v.2)
r+ and fingers crossed - try looked also ok
Attachment #8763826 -
Flags: review?(cbook) → review+
Comment 181•8 years ago
|
||
Pushed by jmaher@mozilla.com:
https://hg.mozilla.org/integration/mozilla-inbound/rev/bd33dc6449d7
90 minute timeout for linux64 mochitest-browser-chrome chunks. r=Tomcat
Comment 182•8 years ago
|
||
bugherder |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 185•8 years ago
|
||
:gbrown, looking at orangefactor in the previous comment, the majority of the 29 failures are linux64 asan tests, primarily:
* [TC] Linux64 mochitest-media-e10s
* [TC] Linux64 xpcshell-6
can you investigate those and fix the timeouts or split up the tests accordingly?
Flags: needinfo?(gbrown)
Comment 186•8 years ago
|
||
Pushed by gbrown@mozilla.com:
https://hg.mozilla.org/integration/mozilla-inbound/rev/5c0b7eae936a
Adjust chunks and maxRunTime to avoid tc Linux x86 intermittent timeouts; r=me
Comment hidden (Intermittent Failures Robot) |
Comment 189•8 years ago
|
||
bugherder |
Comment hidden (Intermittent Failures Robot) |
Reporter | ||
Updated•8 years ago
|
Summary: Intermittent [taskcluster:error] Task timeout after 3600 seconds. Force killing container. / [taskcluster:error] Task timeout after 5400 seconds. Force killing container. → Intermittent [taskcluster:error] Task timeout after 3600 seconds. Force killing container. / [taskcluster:error] Task timeout after 5400 seconds. Force killing container. / [taskcluster:error] Task timeout after 7200 seconds. Force killing container.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 193•8 years ago
|
||
Pushed by gbrown@mozilla.com:
https://hg.mozilla.org/integration/mozilla-inbound/rev/2584ac065137
Increase Android maxRunTime to avoid timeouts; r=me
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 196•8 years ago
|
||
bugherder |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 199•8 years ago
|
||
oh fun, of the 35 posted results for yesterday this is a random mix of jobs, which means there is probably no easy win to fix stuff here. What we need to do is fix the error messages in general.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 206•8 years ago
|
||
Android mochitest-chrome timeouts dominate recent reports here; those are being addressed in bug 1287455.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 213•8 years ago
|
||
Pushed by gbrown@mozilla.com:
https://hg.mozilla.org/integration/mozilla-inbound/rev/6a6829ccc2b4
Adjust chunks and maxRunTime to avoid tc Android mochitest-media and xpcshell timeouts; r=me
Comment hidden (Intermittent Failures Robot) |
Comment 215•8 years ago
|
||
bugherder |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 223•8 years ago
|
||
:aki, can you look into these funsize issues? It was said back in February that you were doing a new funsize worker and that would fix most of this problem.
Flags: needinfo?(aki)
Comment 224•8 years ago
|
||
(In reply to Rail Aliiev [:rail] from comment #119)
> I looked at those and most of them are timeout due to balrog submission
> retries. This is a known issue and aki is going to look at new worker type
> for balrog submission.
It looks like this is still the case. I think this is the same issue being worked on in bug 1284516, but maybe bhearsum can give more information.
Component: General → General Automation
Flags: needinfo?(aki) → needinfo?(bhearsum)
Product: Taskcluster → Release Engineering
QA Contact: catlee
Comment 225•8 years ago
|
||
(In reply to Joel Maher ( :jmaher) from comment #223)
> :aki, can you look into these funsize issues? It was said back in February
> that you were doing a new funsize worker and that would fix most of this
> problem.
This would be Balrog Worker, which isn't done yet. https://bugzilla.mozilla.org/show_bug.cgi?id=1277871 was tracking that work, I'm not sure of the current status though.
(In reply to Dustin J. Mitchell [:dustin] from comment #224)
> (In reply to Rail Aliiev [:rail] from comment #119)
> > I looked at those and most of them are timeout due to balrog submission
> > retries. This is a known issue and aki is going to look at new worker type
> > for balrog submission.
>
> It looks like this is still the case. I think this is the same issue being
> worked on in bug 1284516, but maybe bhearsum can give more information.
Yep, all still true. We're trying a couple of things to mitigate in bug 1284516.
Flags: needinfo?(bhearsum)
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 228•8 years ago
|
||
Pushed by gbrown@mozilla.com:
https://hg.mozilla.org/integration/mozilla-inbound/rev/a957d173402b
Adjust chunks and max run time for Android mochitests; r=me
Comment 229•8 years ago
|
||
bugherder |
Reporter | ||
Comment 230•8 years ago
|
||
Summary: Intermittent [taskcluster:error] Task timeout after 3600 seconds. Force killing container. / [taskcluster:error] Task timeout after 5400 seconds. Force killing container. / [taskcluster:error] Task timeout after 7200 seconds. Force killing container. → Intermittent [taskcluster:error] Task timeout after 3600, 5400, 7200, 10800 seconds. Force killing container.
Reporter | ||
Comment 231•8 years ago
|
||
Fine, that doesn't work, let's see if the word 'or' works around treeherder's broken search.
Summary: Intermittent [taskcluster:error] Task timeout after 3600, 5400, 7200, 10800 seconds. Force killing container. → Intermittent [taskcluster:error] Task timeout after 3600 or 5400 or 7200 or 10800 seconds. Force killing container.
Reporter | ||
Updated•8 years ago
|
Summary: Intermittent [taskcluster:error] Task timeout after 3600 or 5400 or 7200 or 10800 seconds. Force killing container. → Intermittent [taskcluster:error] Task timeout after 3600 seconds. Force killing container. / [taskcluster:error] Task timeout after 5400 seconds. Force killing container. / [taskcluster:error] Task timeout after 7200 seconds. Force killing container.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 236•8 years ago
|
||
Pushed by gbrown@mozilla.com:
https://hg.mozilla.org/integration/mozilla-inbound/rev/971940ade414
Adjust chunks for Android xpcshell tests to avoid intermittent timeouts; r=me
Reporter | ||
Comment 237•8 years ago
|
||
bugherder |
Comment hidden (Intermittent Failures Robot) |
Comment 276•8 years ago
|
||
bugherder uplift |
https://hg.mozilla.org/releases/mozilla-aurora/rev/02b4761289f0
https://hg.mozilla.org/releases/mozilla-aurora/rev/4af3cec722c0
status-firefox51:
--- → fixed
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 282•8 years ago
|
||
something odd happened here on November 8th/9th (I think the 9th) and bc3/bc4 on linux64 asan e10s is taking longer. Possibly investigating times before/after the 9th and seeing if we are right at the limit for 3600 seconds. Either more chunks or longer timeout- I prefer the more chunks approach.
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Updated•8 years ago
|
Assignee: nobody → gbrown
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 299•8 years ago
|
||
About half of the recent failures here are Android; I suspect those are a consequence of bug 1321605. Selecting, for instance, a recent mozilla-central Android Debug crashtest-5 timeout after 60 minutes and comparing to recent mozilla-central successful runs of Android crashtest-5, I see those complete in 30 to 40 minutes. Unfortunately, these timeout failures do not include all the artifacts, so the recently added android-performance.log is missing from these failures...I'll try to sort out bug 1321605 some other way.
Assignee | ||
Comment 300•8 years ago
|
||
Recent Linux failures are more consistent, always in test-linux64/debug-mochitest-media-e10s, and ending in:
[task 2017-02-11T00:45:42.283184Z] 00:45:42 INFO - [Child 5240] WARNING: MsgDropped in ContentChild: file /home/worker/workspace/build/src/dom/ipc/ContentChild.cpp, line 2049
[task 2017-02-11T00:45:42.283380Z] 00:45:42 INFO - [Child 5240] WARNING: '!contentChild->SendAccumulateChildKeyedHistogram(keyedAccumulationsToSend)', file /home/worker/workspace/build/src/toolkit/components/telemetry/TelemetryIPCAccumulator.cpp, line 215
[task 2017-02-11T00:45:44.284873Z] 00:45:44 INFO - ###!!! [Child][MessageChannel] Error: (msgtype=0x4400FD,name=PContent::Msg_AccumulateChildKeyedHistogram) Closed channel: cannot send/recv
[task 2017-02-11T00:45:44.286157Z] 00:45:44 INFO - [Child 5240] WARNING: MsgDropped in ContentChild: file /home/worker/workspace/build/src/dom/ipc/ContentChild.cpp, line 2049
[task 2017-02-11T00:45:44.286317Z] 00:45:44 INFO - [Child 5240] WARNING: '!contentChild->SendAccumulateChildKeyedHistogram(keyedAccumulationsToSend)', file /home/worker/workspace/build/src/toolkit/components/telemetry/TelemetryIPCAccumulator.cpp, line 215
[taskcluster:error] Task timeout after 5400 seconds. Force killing container.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 306•8 years ago
|
||
I believe we have documented the main failures in the related bugs. Ideally within the next week we should see a reduction.
Updated•8 years ago
|
Whiteboard: [stockwell needswork]
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 309•8 years ago
|
||
I believe the work to reduce this a bit is still in bug 1341466- waiting on a fix there.
Comment hidden (Intermittent Failures Robot) |
Comment 311•8 years ago
|
||
the android failures are reduced (almost 100%), but linux64 is spiked a bit.
linux* debug mochitest-media-e10s-2
linux64 asan browser-chrome-e10s-4
lets see if this continues the same pattern, then look into it
Assignee | ||
Comment 312•8 years ago
|
||
Linux mochitest-media failures are very likely bug 1339568.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 315•8 years ago
|
||
there is a mix of mochitest-media-e10s-2 and mochitest-browser-chrome for linux; maybe we could do some try pushes and try to reduce the failures in mochitest-media-e10s-2 to determine which test (or set of tests) is causing this crash? this doesn't appear in mochitest-media-e10s-1, so I am optimistic this can be narrowed down a bit more.
Flags: needinfo?(gbrown)
Assignee | ||
Comment 316•8 years ago
|
||
I agree there probably is some test or set of tests that is "causing" the shutdown hangs in mochitest-media-e10s-2, and bisection on try should be able to find it...but it will require a lot of retries. We rarely see more than 20 such shutdown hangs in a week, and those are evenly distributed across linux-debug, linux64-debug, and linux64-asan, each of which run 100+ times per week - so maybe 1 failure in 20 on one of those platforms. The last time I tried, I could not reproduce the shutdown hang on try at all (probably just bad luck).
Flags: needinfo?(gbrown)
Comment 317•8 years ago
|
||
maybe splitting this into 4 chunks for linux would help reduce the scope here?
Assignee | ||
Comment 318•8 years ago
|
||
See https://bugzilla.mozilla.org/show_bug.cgi?id=1339568#c18 - mochitest-media shutdown hang possibly isolated to about 50 tests.
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 320•8 years ago
|
||
I reviewed recent linux asan e10s mochitest-bc failures, which timed out after 3600 seconds. Those jobs were progressing before the time out - as though they would succeed with more time. I didn't see errors or temporary hangs in the logs. And yet, the same jobs *normally* run in 45 minutes or less -- I don't think I can justify increasing chunks or max time.
Comment 321•8 years ago
|
||
is it possible that we have more overhead in setting up and then taskcluster is timing out; so if we spend 18 minutes setting up docker, and 42 minutes running tests, then we cross the threshold; if docker is already setup, then we spend 1 minute with that setup and 42 minutes testing- much faster.
Assignee | ||
Comment 322•8 years ago
|
||
Setting up the docker image does take a lot of time, but it seems that the 3600 second clock doesn't start ticking until after that's complete.
https://public-artifacts.taskcluster.net/WAcsEzGqQwu2xqkgJUmjeg/0/public/logs/live_backing.log
[taskcluster 2017-04-05 22:02:15.355Z] Decompressing downloaded image
[taskcluster 2017-04-05 22:05:36.493Z] Loading docker image from downloaded archive.
[taskcluster 2017-04-05 22:29:37.849Z] Image 'public/image.tar.zst' from task 'K_S2d2yZTUidRxtsaRZiyA' loaded. Using image ID sha256:2439f499c56d78844c3f8ff1166c772f9a3d7ee0f21b205483be73388698b37e.
[taskcluster 2017-04-05 22:29:38.270Z] === Task Starting ===
...
[taskcluster:error] Task timeout after 3600 seconds. Force killing container.
[taskcluster 2017-04-05 23:29:39.088Z] === Task Finished ===
[taskcluster 2017-04-05 23:29:39.090Z] Unsuccessful task run with exit code: -1 completed in 5391.512 seconds
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 325•8 years ago
|
||
looking at the data for the last week, I see:
8 - linux64 debug mochitest-media-e10s-2
7 - linux64 asan mochitest-media-e10s-2
6 - linux32 debug mochitest-media-e10s-2
10 - linux64 asan bc-e10s-* (2,4,5,6,8)
randoms:
1 - linux64 debug bc-e10s-16
1 - linux64 debug mochitest-3
1 - linux64 qr debug reftest-e10s-6
1 - linux64 asan bc-2 (non e10s)
1 - linux32 debug xpcshell-8
1 - linux32 debug bc-15
1 - linux32 debug mochitest-9
13 - android debug *
1 - linux64 asan xpcshell-8
2 - linux64-stylo builds
21 of the 53 are media-e10s-2, that is sizeable. Looking at this log:
https://public-artifacts.taskcluster.net/VieHxCIJTAGcphPEW__VWQ/0/public/logs/live_backing.log
I see 14 minutes into the task:
task 2017-04-16T03:44:21.424292Z] 03:44:21 INFO - GECKO(2762) | Hit MOZ_CRASH(Shutdown too long, probably frozen, causing a crash.) at /home/worker/workspace/build/src/toolkit/components/terminator/nsTerminator.cpp:159
[task 2017-04-16T03:44:21.508232Z] 03:44:21 INFO - GECKO(2762) | #01: _pt_root [nsprpub/pr/src/pthreads/ptthread.c:219]
[task 2017-04-16T03:44:21.508326Z] 03:44:21 INFO -
[task 2017-04-16T03:44:21.509149Z] 03:44:21 INFO - GECKO(2762) | #02: libpthread.so.0 + 0x76ba
[task 2017-04-16T03:44:21.509304Z] 03:44:21 INFO -
[task 2017-04-16T03:44:21.509355Z] 03:44:21 INFO - GECKO(2762) | #03: libc.so.6 + 0x10682d
[task 2017-04-16T03:44:21.509389Z] 03:44:21 INFO -
[task 2017-04-16T03:44:21.509715Z] 03:44:21 INFO - GECKO(2762) | #04: ??? (???:???)
[task 2017-04-16T03:44:21.509805Z] 03:44:21 INFO - GECKO(2762) | ExceptionHandler::GenerateDump cloned child 3006
this is after finishing the tests in /tests/dom/media/tests/mochitest/identity/ and we were shutting down and cycling the browser as we do between all directories.
I see very similar patterns in many of the other linux64-debug media-e10s-2 logs, except it isn't always the identity directory.
looking at linux64-asan, I see a similar pattern- but more data, for example in this log:
https://treeherder.mozilla.org/logviewer.html#?repo=autoland&job_id=90372228&lineNumber=31892
I see:
[task 2017-04-11T07:18:26.817173Z] 07:18:26 INFO - GECKO(3229) | ASAN:DEADLYSIGNAL
[task 2017-04-11T07:18:26.818010Z] 07:18:26 INFO - GECKO(3229) | =================================================================
[task 2017-04-11T07:18:26.819006Z] 07:18:26 INFO - GECKO(3229) | ==3229==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000000 (pc 0x7fbddb0eb3de bp 0x7fbd9afaed70 sp 0x7fbd9afaed60 T346)
[task 2017-04-11T07:18:26.819829Z] 07:18:26 INFO - GECKO(3229) | ==3229==The signal is caused by a WRITE memory access.
[task 2017-04-11T07:18:26.820613Z] 07:18:26 INFO - GECKO(3229) | ==3229==Hint: address points to the zero page.
[task 2017-04-11T07:18:27.039049Z] 07:18:27 INFO - GECKO(3229) | ###!!! [Child][MessageChannel] Error: (msgtype=0x4800FA,name=PContent::Msg_AccumulateChildKeyedHistograms) Closed channel: cannot send/recv
[task 2017-04-11T07:18:27.225624Z] 07:18:27 INFO - GECKO(3229) | #0 0x7fbddb0eb3dd in mozilla::(anonymous namespace)::RunWatchdog(void*) /home/worker/workspace/build/src/toolkit/components/terminator/nsTerminator.cpp:159:5
[task 2017-04-11T07:18:27.227035Z] 07:18:27 INFO - GECKO(3229) | #1 0x7fbde75d8c93 in _pt_root /home/worker/workspace/build/src/nsprpub/pr/src/pthreads/ptthread.c:216:5
[task 2017-04-11T07:18:27.231239Z] 07:18:27 INFO - GECKO(3229) | #2 0x7fbdeb849e99 in start_thread /build/eglibc-FTTGU2/eglibc-2.15/nptl/pthread_create.c:308
[task 2017-04-11T07:18:27.273393Z] 07:18:27 INFO - GECKO(3229) | #3 0x7fbdea9452ec in clone /build/eglibc-FTTGU2/eglibc-2.15/misc/../sysdeps/unix/sysv/linux/x86_64/clone.S:112
[task 2017-04-11T07:18:27.273481Z] 07:18:27 INFO - GECKO(3229) | AddressSanitizer can not provide additional info.
[task 2017-04-11T07:18:27.273605Z] 07:18:27 INFO - GECKO(3229) | SUMMARY: AddressSanitizer: SEGV /home/worker/workspace/build/src/toolkit/components/terminator/nsTerminator.cpp:159:5 in mozilla::(anonymous namespace)::RunWatchdog(void*)
[task 2017-04-11T07:18:27.273678Z] 07:18:27 INFO - GECKO(3229) | Thread T346 (Shutdow~minator) created by T0 here:
[task 2017-04-11T07:18:27.276187Z] 07:18:27 INFO - GECKO(3229) | #0 0x4a3b76 in __interceptor_pthread_create /builds/slave/moz-toolchain/src/llvm/projects/compiler-rt/lib/asan/asan_interceptors.cc:245:3
[task 2017-04-11T07:18:27.277645Z] 07:18:27 INFO - GECKO(3229) | #1 0x7fbde75d5a39 in _PR_CreateThread /home/worker/workspace/build/src/nsprpub/pr/src/pthreads/ptthread.c:457:14
[task 2017-04-11T07:18:27.279294Z] 07:18:27 INFO - GECKO(3229) | #2 0x7fbde75d564e in PR_CreateThread /home/worker/workspace/build/src/nsprpub/pr/src/pthreads/ptthread.c:548:12
[task 2017-04-11T07:18:27.280508Z] 07:18:27 INFO - GECKO(3229) | #3 0x7fbddb0ebba7 in CreateSystemThread /home/worker/workspace/build/src/toolkit/components/terminator/nsTerminator.cpp:73:22
[task 2017-04-11T07:18:27.281842Z] 07:18:27 INFO - GECKO(3229) | #4 0x7fbddb0ebba7 in StartWatchdog /home/worker/workspace/build/src/toolkit/components/terminator/nsTerminator.cpp:395
[task 2017-04-11T07:18:27.284486Z] 07:18:27 INFO - GECKO(3229) | #5 0x7fbddb0ebba7 in Start /home/worker/workspace/build/src/toolkit/components/terminator/nsTerminator.cpp:359
[task 2017-04-11T07:18:27.286411Z] 07:18:27 INFO - GECKO(3229) | #6 0x7fbddb0ebba7 in mozilla::nsTerminator::Observe(nsISupports*, char const*, char16_t const*) /home/worker/workspace/build/src/toolkit/components/terminator/nsTerminator.cpp:450
[task 2017-04-11T07:18:27.287626Z] 07:18:27 INFO - GECKO(3229) | #7 0x7fbdd19edb1c in nsObserverList::NotifyObservers(nsISupports*, char const*, char16_t const*) /home/worker/workspace/build/src/xpcom/ds/nsObserverList.cpp:112:19
[task 2017-04-11T07:18:27.288976Z] 07:18:27 INFO - GECKO(3229) | #8 0x7fbdd19f1514 in nsObserverService::NotifyObservers(nsISupports*, char const*, char16_t const*) /home/worker/workspace/build/src/xpcom/ds/nsObserverService.cpp:281:19
[task 2017-04-11T07:18:27.290641Z] 07:18:27 INFO - GECKO(3229) | #9 0x7fbddaf59c55 in nsAppStartup::Quit(unsigned int) /home/worker/workspace/build/src/toolkit/components/startup/nsAppStartup.cpp:461:19
[task 2017-04-11T07:18:27.292582Z] 07:18:27 INFO - GECKO(3229) | #10 0x7fbdd1af1ba1 in NS_InvokeByIndex /home/worker/workspace/build/src/xpcom/reflect/xptcall/md/unix/xptcinvoke_asm_x86_64_unix.S:115
[task 2017-04-11T07:18:27.293949Z] 07:18:27 INFO - GECKO(3229) | #11 0x7fbdd31dbd74 in Invoke /home/worker/workspace/build/src/js/xpconnect/src/XPCWrappedNative.cpp:2010:12
[task 2017-04-11T07:18:27.295588Z] 07:18:27 INFO - GECKO(3229) | #12 0x7fbdd31dbd74 in Call /home/worker/workspace/build/src/js/xpconnect/src/XPCWrappedNative.cpp:1329
[task 2017-04-11T07:18:27.297049Z] 07:18:27 INFO - GECKO(3229) | #13 0x7fbdd31dbd74 in XPCWrappedNative::CallMethod(XPCCallContext&, XPCWrappedNative::CallMode) /home/worker/workspace/build/src/js/xpconnect/src/XPCWrappedNative.cpp:1296
[task 2017-04-11T07:18:27.298443Z] 07:18:27 INFO - GECKO(3229) | #14 0x7fbdd31e2f1c in XPC_WN_CallMethod(JSContext*, unsigned int, JS::Value*) /home/worker/workspace/build/src/js/xpconnect/src/XPCWrappedNativeJSOps.cpp:983:12
[task 2017-04-11T07:18:27.299930Z] 07:18:27 INFO - GECKO(3229) | #15 0x3fea13eff1e3 (<unknown module>)
[task 2017-04-11T07:18:27.301257Z] 07:18:27 INFO - GECKO(3229) | #16 0x621001b0684f (<unknown module>)
[task 2017-04-11T07:18:27.302650Z] 07:18:27 INFO - GECKO(3229) | #17 0x3fea13c878a5 (<unknown module>)
[task 2017-04-11T07:18:27.305155Z] 07:18:27 INFO - GECKO(3229) | #18 0x7fbddb81faa2 in EnterBaseline(JSContext*, js::jit::EnterJitData&) /home/worker/workspace/build/src/js/src/jit/BaselineJIT.cpp:160:9
[task 2017-04-11T07:18:27.306743Z] 07:18:27 INFO - GECKO(3229) | #19 0x7fbddb81f307 in js::jit::EnterBaselineMethod(JSContext*, js::RunState&) /home/worker/workspace/build/src/js/src/jit/BaselineJIT.cpp:200:28
[task 2017-04-11T07:18:27.308468Z] 07:18:27 INFO - GECKO(3229) | #20 0x7fbddb5a9165 in js::RunScript(JSContext*, js::RunState&) /home/worker/workspace/build/src/js/src/vm/Interpreter.cpp:385:41
[task 2017-04-11T07:18:27.310046Z] 07:18:27 INFO - GECKO(3229) | #21 0x7fbddb5da5d8 in js::InternalCallOrConstruct(JSContext*, JS::CallArgs const&, js::MaybeConstruct) /home/worker/workspace/build/src/js/src/vm/Interpreter.cpp:473:15
[task 2017-04-11T07:18:27.311189Z] 07:18:27 INFO - GECKO(3229) | #22 0x7fbddb5dae02 in js::Call(JSContext*, JS::Handle<JS::Value>, JS::Handle<JS::Value>, js::AnyInvokeArgs const&, JS::MutableHandle<JS::Value>) /home/worker/workspace/build/src/js/src/vm/Interpreter.cpp:519:10
[task 2017-04-11T07:18:27.312381Z] 07:18:27 INFO - GECKO(3229) | #23 0x7fbddbf5a323 in JS_CallFunctionValue(JSContext*, JS::Handle<JSObject*>, JS::Handle<JS::Value>, JS::HandleValueArray const&, JS::MutableHandle<JS::Value>) /home/worker/workspace/build/src/js/src/jsapi.cpp:2826:12
[task 2017-04-11T07:18:27.314441Z] 07:18:27 INFO - GECKO(3229) | #24 0x7fbdd407c334 in nsFrameMessageManager::ReceiveMessage(nsISupports*, nsIFrameLoader*, bool, nsAString const&, bool, mozilla::dom::ipc::StructuredCloneData*, mozilla::jsipc::CpowHolder*, nsIPrincipal*, nsTArray<mozilla::dom::ipc::StructuredCloneData>*) /home/worker/workspace/build/src/dom/base/nsFrameMessageManager.cpp:1108:14
[task 2017-04-11T07:18:27.316220Z] 07:18:27 INFO - GECKO(3229) | #25 0x7fbdd407d068 in nsFrameMessageManager::ReceiveMessage(nsISupports*, nsIFrameLoader*, bool, nsAString const&, bool, mozilla::dom::ipc::StructuredCloneData*, mozilla::jsipc::CpowHolder*, nsIPrincipal*, nsTArray<mozilla::dom::ipc::StructuredCloneData>*) /home/worker/workspace/build/src/dom/base/nsFrameMessageManager.cpp:1138:29
[task 2017-04-11T07:18:27.317351Z] 07:18:27 INFO - GECKO(3229) | #26 0x7fbdd407d068 in nsFrameMessageManager::ReceiveMessage(nsISupports*, nsIFrameLoader*, bool, nsAString const&, bool, mozilla::dom::ipc::StructuredCloneData*, mozilla::jsipc::CpowHolder*, nsIPrincipal*, nsTArray<mozilla::dom::ipc::StructuredCloneData>*) /home/worker/workspace/build/src/dom/base/nsFrameMessageManager.cpp:1138:29
[task 2017-04-11T07:18:27.318885Z] 07:18:27 INFO - GECKO(3229) | #27 0x7fbdd407d068 in nsFrameMessageManager::ReceiveMessage(nsISupports*, nsIFrameLoader*, bool, nsAString const&, bool, mozilla::dom::ipc::StructuredCloneData*, mozilla::jsipc::CpowHolder*, nsIPrincipal*, nsTArray<mozilla::dom::ipc::StructuredCloneData>*) /home/worker/workspace/build/src/dom/base/nsFrameMessageManager.cpp:1138:29
[task 2017-04-11T07:18:27.320432Z] 07:18:27 INFO - GECKO(3229) | #28 0x7fbdd4079a79 in nsFrameMessageManager::ReceiveMessage(nsISupports*, nsIFrameLoader*, nsAString const&, bool, mozilla::dom::ipc::StructuredCloneData*, mozilla::jsipc::CpowHolder*, nsIPrincipal*, nsTArray<mozilla::dom::ipc::StructuredCloneData>*) /home/worker/workspace/build/src/dom/base/nsFrameMessageManager.cpp:917:10
[task 2017-04-11T07:18:27.321941Z] 07:18:27 INFO - GECKO(3229) | #29 0x7fbdd745ba7d in mozilla::dom::TabParent::ReceiveMessage(nsString const&, bool, mozilla::dom::ipc::StructuredCloneData*, mozilla::jsipc::CpowHolder*, nsIPrincipal*, nsTArray<mozilla::dom::ipc::StructuredCloneData>*) /home/worker/workspace/build/src/dom/ipc/TabParent.cpp:2414:14
[task 2017-04-11T07:18:27.323074Z] 07:18:27 INFO - GECKO(3229) | #30 0x7fbdd746a31f in mozilla::dom::TabParent::RecvAsyncMessage(nsString const&, nsTArray<mozilla::jsipc::CpowEntry>&&, IPC::Principal const&, mozilla::dom::ClonedMessageData const&) /home/worker/workspace/build/src/dom/ipc/TabParent.cpp:1607:8
[task 2017-04-11T07:18:27.324261Z] 07:18:27 INFO - GECKO(3229) | #31 0x7fbdd2ddc871 in mozilla::dom::PBrowserParent::OnMessageReceived(IPC::Message const&) /home/worker/workspace/build/src/obj-firefox/ipc/ipdl/PBrowserParent.cpp:1644:20
[task 2017-04-11T07:18:27.327795Z] 07:18:27 INFO - GECKO(3229) | #32 0x7fbdd2f555d3 in mozilla::dom::PContentParent::OnMessageReceived(IPC::Message const&) /home/worker/workspace/build/src/obj-firefox/ipc/ipdl/PContentParent.cpp:3083:28
[task 2017-04-11T07:18:27.328916Z] 07:18:27 INFO - GECKO(3229) | #33 0x7fbdd288deb0 in mozilla::ipc::MessageChannel::DispatchAsyncMessage(IPC::Message const&) /home/worker/workspace/build/src/ipc/glue/MessageChannel.cpp:1872:25
[task 2017-04-11T07:18:27.329960Z] 07:18:27 INFO - GECKO(3229) | #34 0x7fbdd288a6f7 in mozilla::ipc::MessageChannel::DispatchMessage(IPC::Message&&) /home/worker/workspace/build/src/ipc/glue/MessageChannel.cpp:1807:17
[task 2017-04-11T07:18:27.331106Z] 07:18:27 INFO - GECKO(3229) | #35 0x7fbdd288cb24 in mozilla::ipc::MessageChannel::RunMessage(mozilla::ipc::MessageChannel::MessageTask&) /home/worker/workspace/build/src/ipc/glue/MessageChannel.cpp:1680:5
[task 2017-04-11T07:18:27.332211Z] 07:18:27 INFO - GECKO(3229) | #36 0x7fbdd288d126 in mozilla::ipc::MessageChannel::MessageTask::Run() /home/worker/workspace/build/src/ipc/glue/MessageChannel.cpp:1713:15
[task 2017-04-11T07:18:27.333689Z] 07:18:27 INFO - GECKO(3229) | #37 0x7fbdd1ad7410 in nsThread::ProcessNextEvent(bool, bool*) /home/worker/workspace/build/src/xpcom/threads/nsThread.cpp:1269:14
[task 2017-04-11T07:18:27.335265Z] 07:18:27 INFO - GECKO(3229) | #38 0x7fbdd1af1ba1 in NS_InvokeByIndex /home/worker/workspace/build/src/xpcom/reflect/xptcall/md/unix/xptcinvoke_asm_x86_64_unix.S:115
[task 2017-04-11T07:18:27.336778Z] 07:18:27 INFO - GECKO(3229) | #39 0x7fbdd31dbd74 in Invoke /home/worker/workspace/build/src/js/xpconnect/src/XPCWrappedNative.cpp:2010:12
[task 2017-04-11T07:18:27.338161Z] 07:18:27 INFO - GECKO(3229) | #40 0x7fbdd31dbd74 in Call /home/worker/workspace/build/src/js/xpconnect/src/XPCWrappedNative.cpp:1329
[task 2017-04-11T07:18:27.339239Z] 07:18:27 INFO - GECKO(3229) | #41 0x7fbdd31dbd74 in XPCWrappedNative::CallMethod(XPCCallContext&, XPCWrappedNative::CallMode) /home/worker/workspace/build/src/js/xpconnect/src/XPCWrappedNative.cpp:1296
looking in more detail about the media-e10s-2 failures, on asan, I looked at 2 logs:
pass: https://public-artifacts.taskcluster.net/aj3W_tAaRnq5qVkCanO7uA/0/public/logs/live_backing.log
fail: https://public-artifacts.taskcluster.net/fKq9ixPMQvO-H6r9uuAdYw/0/public/logs/live_backing.log
there is a 45 second difference from when the browser starts up to when we run browser_closeTabSpecificPanels.js. The failure case has more of these, in fact 506 seconds of slower runTime overall- this would account for us timing out.
while this doesn't account for everything, should we increase the total runtime and/or debug why startup is taking so long?
Assignee | ||
Comment 326•8 years ago
|
||
The nsTerminator crash is discussed in bug 1339568.
Assignee | ||
Comment 327•8 years ago
|
||
(In reply to Joel Maher ( :jmaher) from comment #325)
> looking in more detail about the media-e10s-2 failures, on asan, I looked at 2 logs:
It looks like those logs are from mochitest-bc-e10s-5 jobs. Right?
> there is a 45 second difference from when the browser starts up to when we
> run browser_closeTabSpecificPanels.js. The failure case has more of these,
> in fact 506 seconds of slower runTime overall- this would account for us
> timing out.
I see the 45 seconds, but startup often takes 20 seconds anyway on a "good" run, I'd say, so ~ +25 seconds? I see that.
I'm not sure I see the 506 seconds...but if you do, that seems interesting.
> while this doesn't account for everything, should we increase the total
> runtime and/or debug why startup is taking so long?
It looks like some linux64-asan mochitest-bc chunks, especially e10s ones, are running a little long...longer than I recall from my check in comment 320. Maybe more chunks are in order?
Comment 328•8 years ago
|
||
we have 16 chunks on linux debug, we should go from 10 to 16 for asan :)
Assignee | ||
Comment 329•8 years ago
|
||
Attachment #8859398 -
Flags: review?(jmaher)
Comment 330•8 years ago
|
||
Comment on attachment 8859398 [details] [diff] [review]
increase linux64-asan mochitest-bc chunks from 10 to 16
Review of attachment 8859398 [details] [diff] [review]:
-----------------------------------------------------------------
nice an simple
Attachment #8859398 -
Flags: review?(jmaher) → review+
Comment 331•8 years ago
|
||
Pushed by gbrown@mozilla.com:
https://hg.mozilla.org/integration/mozilla-inbound/rev/72abcde6295f
Increase number of test chunks for linux64-asan mochitest-bc; r=jmaher
Comment 332•8 years ago
|
||
backed out since i guess this caused https://treeherder.mozilla.org/logviewer.html#?job_id=92712773&repo=mozilla-inbound because it seems this only affects asan
Flags: needinfo?(gbrown)
Comment 333•8 years ago
|
||
Backout by cbook@mozilla.com:
https://hg.mozilla.org/integration/mozilla-inbound/rev/f1c0c2410568
Backed out changeset 72abcde6295f for suspicion this cause asan test failures
Assignee | ||
Comment 334•8 years ago
|
||
Failures persisted after my backout: see https://bugzilla.mozilla.org/show_bug.cgi?id=867815#c12.
Flags: needinfo?(gbrown)
Comment 335•8 years ago
|
||
Pushed by gbrown@mozilla.com:
https://hg.mozilla.org/integration/mozilla-inbound/rev/21b982b24bd5
Increase number of test chunks for linux64-asan mochitest-bc; r=jmaher
Comment 336•8 years ago
|
||
bugherder |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 338•8 years ago
|
||
No new asan mochitest-bc failures - it looks like the new chunks worked.
Bug 1339568 continues, accounting for about 50% of recent failures.
Comment hidden (Intermittent Failures Robot) |
Comment 340•8 years ago
|
||
uplift |
(In reply to Wes Kocher (:KWierso) from comment #336)
> https://hg.mozilla.org/mozilla-central/rev/21b982b24bd5
And to Beta.
https://hg.mozilla.org/releases/mozilla-beta/rev/23228e9d57e3
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 342•8 years ago
|
||
Bug 1339568 continues, accounting for about 50% of recent failures.
There is a new problem which may be actionable: linux32-debug mochitest-6 and mochitest-e10s-6 are running more than twice as long as other chunks, sometimes timing out at 5400 seconds.
Comment 343•8 years ago
|
||
possibly a few tests or a set of tests are much longer than before? That could be actionable in addition to more chunks.
Assignee | ||
Comment 344•8 years ago
|
||
https://public-artifacts.taskcluster.net/O0q4N3GuSsqoNBSZRWTV6g/0/public/logs/live_backing.log
00:47:55 INFO - SUITE-START | Running 1589 tests
00:49:36 INFO - Slowest: 4874ms - /tests/dom/tests/mochitest/dom-level0/test_innerWidthHeight_script.html
01:09:15 INFO - Slowest: 4894ms - /tests/dom/tests/mochitest/dom-level1-core/test_PIsetdatanomodificationallowederrEE.html
01:17:55 INFO - Slowest: 3974ms - /tests/dom/tests/mochitest/dom-level2-core/test_attrgetownerelement01.html
01:43:16 INFO - Slowest: 4553ms - /tests/dom/tests/mochitest/dom-level2-html/test_HTMLDocument11.html
01:53:00 INFO - Slowest: 140574ms - /tests/dom/tests/mochitest/fetch/test_fetch_cors_sw_reroute.html
01:54:10 INFO - Slowest: 3727ms - /tests/dom/tests/mochitest/gamepad/test_check_timestamp.html
01:58:28 INFO - Slowest: 11641ms - /tests/dom/tests/mochitest/general/test_storagePermissionsAccept.html
01:59:24 INFO - Slowest: 8476ms - /tests/dom/tests/mochitest/general/test_interfaces_secureContext.html
02:00:23 INFO - Slowest: 3702ms - /tests/dom/tests/mochitest/geolocation/test_geoGetCurrentPositionBlockedInInsecureContext.html
02:02:52 INFO - Slowest: 21787ms - /tests/dom/tests/mochitest/geolocation/test_manyCurrentSerial.html
02:05:16 INFO - Slowest: 11120ms - /tests/dom/tests/mochitest/localstorage/test_localStorageReplace.html
02:06:13 INFO - Slowest: 3289ms - /tests/dom/tests/mochitest/notification/test_bug931307.html
02:07:07 INFO - Slowest: 4003ms - /tests/dom/tests/mochitest/orientation/test_bug507902.html
02:09:52 INFO - Slowest: 109059ms - /tests/dom/tests/mochitest/pointerlock/test_pointerlock-api.html
02:11:25 INFO - Slowest: 8807ms - /tests/dom/tests/mochitest/sessionstorage/test_sessionStorageReplace.html
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 348•8 years ago
|
||
Huge spike on autoland today for linux64-pgo and opt was fixed by commit https://hg.mozilla.org/integration/autoland/rev/7308157309aebd6a8a889adc70298adec7bd5691 (backout bug 1364068).
Assignee | ||
Comment 349•8 years ago
|
||
(In reply to OrangeFactor Robot from comment #347)
> 28 failures in 147 pushes (0.19 failures/push) were associated with this bug
> yesterday.
>
> Repository breakdown:
> * mozilla-inbound: 12
> * autoland: 10
> * mozilla-central: 5
> * mozilla-beta: 1
>
> Platform breakdown:
> * android-4-3-armv7-api15: 20
> * linux32: 4
> * android-api-15-gradle: 2
> * linux64: 1
> * android-4-2-x86: 1
Many of these Android failures were due to repeated tooltool timeouts like:
https://treeherder.mozilla.org/logviewer.html#?repo=mozilla-beta&job_id=100941800&lineNumber=672
[task 2017-05-22T14:47:55.831708Z] 14:47:55 INFO - Calling ['/usr/bin/python2.7', '/home/worker/workspace/build/tooltool.py', '--url', 'http://relengapi/tooltool/', 'fetch', '-m', '/home/worker/workspace/build/.android/releng.manifest', '-o', '-c', '/home/worker/tooltool_cache'] with output_timeout 600
[task 2017-05-22T14:47:55.880278Z] 14:47:55 INFO - INFO - File AVDs-armv7a-android-4.3.1_r1-build-2016-08-02.tar.gz not present in local cache folder /home/worker/tooltool_cache
[task 2017-05-22T14:47:55.881029Z] 14:47:55 INFO - INFO - Attempting to fetch from 'http://relengapi/tooltool/'...
[task 2017-05-22T14:47:58.807712Z] compiz (core) - Warn: Attempted to restack relative to 0x1600006 which is not a child of the root window or a window compiz owns
[task 2017-05-22T14:57:55.900272Z] 14:57:55 INFO - Automation Error: mozprocess timed out after 600 seconds running ['/usr/bin/python2.7', '/home/worker/workspace/build/tooltool.py', '--url', 'http://relengapi/tooltool/', 'fetch', '-m', '/home/worker/workspace/build/.android/releng.manifest', '-o', '-c', '/home/worker/tooltool_cache']
[task 2017-05-22T14:57:55.906063Z] 14:57:55 ERROR - timed out after 600 seconds of no output
[task 2017-05-22T14:57:55.906387Z] 14:57:55 ERROR - Return code: -9
[task 2017-05-22T14:57:55.906875Z] 14:57:55 INFO - retry: Failed, sleeping 60 seconds before retrying
...presumably a temporary tooltool issue.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 353•8 years ago
|
||
(In reply to OrangeFactor Robot from comment #352)
> 21 failures in 167 pushes (0.126 failures/push) were associated with this
> bug yesterday.
Several of these are test-android-4.3-arm7-api-15/debug-marionette-4. That job doubled in time in this range - I'll try to narrow that down.
https://treeherder.mozilla.org/#/jobs?repo=autoland&filter-searchStr=android%20marionette&tochange=0344daf0fe0cd3903aa872b22f6820e8c40b1b56&fromchange=dec37391ecf8c26962fede1c15db9b5f8c769b28
Flags: needinfo?(gbrown)
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 357•7 years ago
|
||
(In reply to OrangeFactor Robot from comment #356)
> 71 failures in 820 pushes (0.087 failures/push) were associated with this
> bug in the last 7 days.
About 32 of these are bug 1339568 (linux mochitest-media shutdown hang).
About 23 of these are bug 1369083 (android marionette).
There are a few linux-debug mochitest-6 cases, as in comment 342.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 362•7 years ago
|
||
Something has gone wrong in linux64-ccov...need to investigate.
Flags: needinfo?(gbrown)
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Updated•7 years ago
|
Flags: needinfo?(gbrown)
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
The linux64-qr spike in debug R-e10s-5 and Ru-e10s-5 seem to be because the log is filled with webrender warnings, the same warnings I noted in bug 1206887 comment 69. Hopefully fixing that will kill this spike. Also interesting is that this spike started more recently than the APZ enabling. In fact it seems to have started when stylo started getting built by default. I'm doing some retriggers on the range to verify, because one of those patches got backed out and relanded so it might give an extra clue.
https://treeherder.mozilla.org/#/jobs?repo=autoland&filter-searchStr=qr%20reftest%20e10s&tochange=28585cf7da6fdc07ac775ea47ad3aa8fae406351&fromchange=1990807be52407bdba9d61d1883300185c8b9952&group_state=expanded
Assignee | ||
Comment 367•7 years ago
|
||
Thanks kats! I have been trying to sort that out in bug 1375550, but not having much luck. My conclusion was that the warnings and increase in time started with https://treeherder.mozilla.org/#/jobs?repo=autoland&revision=471a163b37d092fc5bf7a56bcf5c5295f727b8d8.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 373•7 years ago
|
||
this has a recent spike, not sure what to do here, but will revisit if this continues with more failures through the weekend.
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 375•7 years ago
|
||
Android and Stylo crashtests are timing out this week.
Flags: needinfo?(gbrown)
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 377•7 years ago
|
||
(In reply to Geoff Brown [:gbrown] from comment #375)
> Android and Stylo crashtests are timing out this week.
Linux (including Stylo) crashtests are under investigation in bug 1381283; Android crashtests in bug 1381839.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 386•7 years ago
|
||
Almost all current failures are bug 1339568.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 390•7 years ago
|
||
Almost all current failures are bug 1339568.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 395•7 years ago
|
||
recently we have had a large spike in linux32*/debug failures, looking like timeouts.
I looked over runtimes, they are all over the map from 16 minutes to 100+ minutes for both mochitest-plain and browser-chrome on linux32-debug (gecko||stylo). if we could balance things out and kept it at 16 chunks we would be between 50-60 minutes for each chunk of mochitest-plain and 75 minutes for each of the 16 browser-chrome chunks.
I then ask:
* why we need to run these so often? (maybe seta for linux*-stylo is needed on autoland?)
* can we balance our chunks out more to have fewer failures
* should we analyze which tests are taking long and see if there are efficiencies we can gain?
* do we need all these tests on linux32 debug?
* should we increase the timeout or the total number of chunks?
:ahal, do you have any thoughts on this subject?
Flags: needinfo?(ahalberstadt)
Comment 396•7 years ago
|
||
Yeah, browser-chrome already has --chunk-by-runtime enabled, but I think the data files haven't been updated in about 6 months or so. So updating those might help, I'll file a bug. For mochitest-plain, --chunk-by-runtime is currently disabled, but we could definitely look into enabling it there too.
I'd also be in favour of increasing the timeout if we're hitting it legitimately. Maybe try balancing things out better first. Note: seta would reduce the absolute number of failures here, but wouldn't help reduce the failure rate.
Flags: needinfo?(ahalberstadt)
Assignee | ||
Comment 397•7 years ago
|
||
linux32-stylo tests were enabled in bug 1385025; timeouts started almost immediately.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment 400•7 years ago
|
||
:gbrown, should we push on getting stylo tests setup on SETA to reduce the frequency?
Comment hidden (Intermittent Failures Robot) |
Comment 402•7 years ago
|
||
In bug 1393900 I updated the runtimes files and did a try push:
https://treeherder.mozilla.org/#/jobs?repo=try&revision=5d37bb3c580eabf0476fe20e1c619e91e6d09ca0
The push looks good, the chunks seem much more balanced. But the maximum runtime (linux debug bc15) is still at ~82 min. If you look at the errorsummary for that job, you'll notice it's only running a single manifest:
https://dxr.mozilla.org/mozilla-central/source/browser/components/extensions/test/browser/browser-common.ini
Because --chunk-by-runtime doesn't split up manifests this means we are already at the smallest atomic chunk for this particular job and adding more chunks won't lower this max runtime by anything. If 82 min is still too long, we can either:
1) Split that manifest in half
2) Increase the timeout
It's possible there's something going wrong in those tests (like a timeout in each test or something) that's causing them to take so long. So fixing the tests might be a 3rd option.
Joel, I think we should still update the runtimes files, but we'll need a better strategy for dealing with cases when a single manifest exceeds the timeout.
Flags: needinfo?(jmaher)
Comment 403•7 years ago
|
||
if we split/fix browser-common.ini, will that require rebalancing again, as in- should we do that before checking in code to rebalance the chunks?
Flags: needinfo?(jmaher) → needinfo?(ahalberstadt)
Comment 404•7 years ago
|
||
No, it won't make a difference, we can do this in either order.
Flags: needinfo?(ahalberstadt)
Comment 405•7 years ago
|
||
ok, then lets go forward with your rebalancing numbers
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 407•7 years ago
|
||
We have had a few Android Mn-1 timeouts reported here over the last couple of weeks; that job was long-running, normally running in 50-60 minutes, with a 60 minute max run time. With changes for bug 1369827, Android Marionette job run times dropped dramatically -- no need for further action.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 410•7 years ago
|
||
Over the last few days, failure rates have dropped, and bug 1339568 dominates (as usual!).
There are a few recent failures in linux32-stylo/debug mochitests...not sure if that's worth follow-up yet.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Updated•7 years ago
|
Flags: needinfo?(gbrown)
Assignee | ||
Updated•7 years ago
|
Flags: needinfo?(gbrown)
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 416•7 years ago
|
||
Failures here are an on-going problem, sometimes related to specific tests, sometimes to specific suites, and sometimes due to test infrastructure problems. I'm marking as "infra" loosely, for convenience.
Keywords: leave-open
Whiteboard: [stockwell disable-recommended] → [stockwell infra]
Assignee | ||
Comment 417•7 years ago
|
||
Some Android reftest failures recorded here are likely "caused" by bug 1401035.
Depends on: 1401035
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Updated•7 years ago
|
Whiteboard: [stockwell infra] → [stockwell needswork]
Assignee | ||
Comment 421•7 years ago
|
||
There was a spike on autoland for timeouts in linux debug jsreftests. It started with bug 1406212, which was backed out for other failures -- seems resolved now.
Bug 1407687 and bug 1339568 remain problematic.
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Comment hidden (Intermittent Failures Robot) |
Assignee | ||
Comment 425•7 years ago
|
||
Dup-ing this to its recent clone since this bug has grown too big and has too much history which is no longer relevant.
Status: NEW → RESOLVED
Closed: 7 years ago
Resolution: --- → DUPLICATE
Comment hidden (Intermittent Failures Robot) |
Updated•7 years ago
|
Whiteboard: [stockwell disable-recommended] → [stockwell infra]
Comment hidden (Intermittent Failures Robot) |
Updated•7 years ago
|
Component: General Automation → General
Comment hidden (Intermittent Failures Robot) |
You need to log in
before you can comment on or make changes to this bug.
Description
•