-
Type:
Bug
-
Resolution: Done
-
Priority:
Major - P3
-
None
-
Affects Version/s: 3.4.0-rc4
-
Component/s: Replication
-
None
-
Replication
-
ALL
-
None
-
None
-
None
-
None
-
None
-
None
-
None
- Start with a 3 node replica set in which all nodes are wiredTiger with journaling disabled
- Make a single write
- Attempt to step down the primary
When attempting the step down command you will get:
backup_test:PRIMARY> rs.stepDown()
{
"ok" : 0,
"errmsg" : "No electable secondaries caught up as of 2016-11-20T21:56:10.466+0000. Please use {force: true} to force node to step down.",
"code" : 50,
"codeName" : "ExceededTimeLimit"
}
The rs.status() looks as below. Not that all the optimeDurable are undefined:
backup_test:PRIMARY> rs.status()
{
"set" : "backup_test",
"date" : ISODate("2016-11-20T21:55:49.329Z"),
"myState" : 1,
"term" : NumberLong(2),
"heartbeatIntervalMillis" : NumberLong(2000),
"optimes" : {
"lastCommittedOpTime" : {
"ts" : Timestamp(0, 0),
"t" : NumberLong(-1)
},
"appliedOpTime" : {
"ts" : Timestamp(1479678946, 1),
"t" : NumberLong(2)
},
"durableOpTime" : {
"ts" : Timestamp(0, 0),
"t" : NumberLong(-1)
}
},
"members" : [
{
"_id" : 0,
"name" : "cailinmac:27000",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 7,
"optime" : {
"ts" : Timestamp(1479678946, 1),
"t" : NumberLong(2)
},
"optimeDurable" : {
"ts" : Timestamp(1479678888, 1),
"t" : NumberLong(1)
},
"optimeDate" : ISODate("2016-11-20T21:55:46Z"),
"optimeDurableDate" : ISODate("2016-11-20T21:54:48Z"),
"lastHeartbeat" : ISODate("2016-11-20T21:55:48.020Z"),
"lastHeartbeatRecv" : ISODate("2016-11-20T21:55:46.518Z"),
"pingMs" : NumberLong(0),
"syncingTo" : "cailinmac:27020",
"configVersion" : 1
},
{
"_id" : 1,
"name" : "cailinmac:27010",
"health" : 1,
"state" : 1,
"stateStr" : "PRIMARY",
"uptime" : 140,
"optime" : {
"ts" : Timestamp(1479678946, 1),
"t" : NumberLong(2)
},
"optimeDate" : ISODate("2016-11-20T21:55:46Z"),
"infoMessage" : "could not find member to sync from",
"electionTime" : Timestamp(1479678905, 1),
"electionDate" : ISODate("2016-11-20T21:55:05Z"),
"configVersion" : 1,
"self" : true
},
{
"_id" : 2,
"name" : "cailinmac:27020",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 120,
"optime" : {
"ts" : Timestamp(1479678946, 1),
"t" : NumberLong(2)
},
"optimeDurable" : {
"ts" : Timestamp(0, 0),
"t" : NumberLong(-1)
},
"optimeDate" : ISODate("2016-11-20T21:55:46Z"),
"optimeDurableDate" : ISODate("1970-01-01T00:00:00Z"),
"lastHeartbeat" : ISODate("2016-11-20T21:55:47.952Z"),
"lastHeartbeatRecv" : ISODate("2016-11-20T21:55:48.912Z"),
"pingMs" : NumberLong(0),
"syncingTo" : "cailinmac:27010",
"configVersion" : 1
}
],
"ok" : 1
}
Here are the command line options for a member. If I enable journaling, this problem disappears.
backup_test:PRIMARY> db.runCommand({getCmdLineOpts:1})
{
"argv" : [
"mongod",
"--port=27000",
"--replSet=backup_test",
"--dbpath=/Users/cailin/Documents/code/mms/data/db/replica/backup_test/backup_test_0",
"--logpath=/Users/cailin/Documents/code/mms/data/db/replica/backup_test/backup_test_0/mongodb.log",
"--logappend",
"--oplogSize=100",
"--storageEngine=wiredTiger",
"--nojournal",
"--wiredTigerEngineConfigString=cache_size=512MB"
],
"parsed" : {
"net" : {
"port" : 27000
},
"replication" : {
"oplogSizeMB" : 100,
"replSet" : "backup_test"
},
"storage" : {
"dbPath" : "/Users/cailin/Documents/code/mms/data/db/replica/backup_test/backup_test_0",
"engine" : "wiredTiger",
"journal" : {
"enabled" : false
},
"wiredTiger" : {
"engineConfig" : {
"configString" : "cache_size=512MB"
}
}
},
"systemLog" : {
"destination" : "file",
"logAppend" : true,
"path" : "/Users/cailin/Documents/code/mms/data/db/replica/backup_test/backup_test_0/mongodb.log"
}
},
"ok" : 1
}
- documents
-
SERVER-27217 Allow nopreallocj to be specified in YAML config file
-
- Closed
-