[SERVER-11177] stale_clustered.js failed on RHEL 5.7 64-bit Created: 14/Oct/13 Updated: 11/Jul/16 Resolved: 17/Oct/13 |
|
| Status: | Closed |
| Project: | Core Server |
| Component/s: | Testing Infrastructure |
| Affects Version/s: | None |
| Fix Version/s: | 2.5.3 |
| Type: | Bug | Priority: | Major - P3 |
| Reporter: | Matt Kangas | Assignee: | Greg Studer |
| Resolution: | Done | Votes: | 0 |
| Labels: | buildbot | ||
| Remaining Estimate: | Not Specified | ||
| Time Spent: | Not Specified | ||
| Original Estimate: | Not Specified | ||
| Backwards Compatibility: | Fully Compatible |
| Operating System: | ALL |
| Participants: |
| Description |
|
Assigning to Greg for triage. a428f58a6fd4 RHEL 5.7 64-bit slow_nightly_tests (logfile)
Last visible green was 6a70c219c629 /cc alerner |
| Comments |
| Comment by Ernie Hershey [ 16/Oct/13 ] | |||||||||||||||||||||||||||||
|
Should I convert /data on that build host to XFS? It's ext4 now, which I think could be contributing to preallocation slowdown. Is the basic build host issue that the filesystem is slow? Or do you think it could be something else? | |||||||||||||||||||||||||||||
| Comment by Matt Kangas [ 16/Oct/13 ] | |||||||||||||||||||||||||||||
|
Ok, sounds like the above (rev 94ce941611a) failures were build host issues. Keeping ticket open until we are clear. | |||||||||||||||||||||||||||||
| Comment by Greg Studer [ 16/Oct/13 ] | |||||||||||||||||||||||||||||
|
The second failure above: > rev 94ce941611a This is a failure to even start up the basic replica set for the test:
and doesn't die...
That particular node, :31201, never actually seems to start accepting connections because it's preallocating.
Things are pretty confusing since the log entries for this node are way out of order, but it looks like it took at least 5 minutes to allocate journal files here, and either we were still allocating or the process itself hung completely. | |||||||||||||||||||||||||||||
| Comment by Matt Kangas [ 16/Oct/13 ] | |||||||||||||||||||||||||||||
|
Two more failures which occurred after commit 2e972e1 rev 94ce941611a
rev 94ce941611a
| |||||||||||||||||||||||||||||
| Comment by auto [ 15/Oct/13 ] | |||||||||||||||||||||||||||||
|
Author: {u'username': u'gregstuder', u'name': u'Greg Studer', u'email': u'greg@10gen.com'}Message: | |||||||||||||||||||||||||||||
| Comment by Daniel Pasette (Inactive) [ 15/Oct/13 ] | |||||||||||||||||||||||||||||
|
also appearing on bb osx and linux-64 | |||||||||||||||||||||||||||||
| Comment by Matt Kangas [ 15/Oct/13 ] | |||||||||||||||||||||||||||||
|
We had a green build at 486efd8f13fc RHEL 5.7 64-bit slow_nightly_tests, followed by another failure. So it's not completely deterministic. 9fcdfeb195c4 C++11 Ubuntu 1204 64-bit slow_nightly_tests also went green but I'm guessing we haven't actually fixed the issue. | |||||||||||||||||||||||||||||
| Comment by Matt Kangas [ 14/Oct/13 ] | |||||||||||||||||||||||||||||
|
Also visible on the C++11 Ubuntu builder. And rather oddly, it seems to 486efd8f13fc C++11 Ubuntu 1204 64-bit slow_nightly_tests Commit before that (ae9897079d21) was green. So the first failed commit was Benety's fix for the Solaris build?. I'm puzzled. /cc benety.goh |